Sample records for text dependent speaker

  1. Left hemisphere lateralization for lexical and acoustic pitch processing in Cantonese speakers as revealed by mismatch negativity.

    PubMed

    Gu, Feng; Zhang, Caicai; Hu, Axu; Zhao, Guoping

    2013-12-01

    For nontonal language speakers, speech processing is lateralized to the left hemisphere and musical processing is lateralized to the right hemisphere (i.e., function-dependent brain asymmetry). On the other hand, acoustic temporal processing is lateralized to the left hemisphere and spectral/pitch processing is lateralized to the right hemisphere (i.e., acoustic-dependent brain asymmetry). In this study, we examine whether the hemispheric lateralization of lexical pitch and acoustic pitch processing in tonal language speakers is consistent with the patterns of function- and acoustic-dependent brain asymmetry in nontonal language speakers. Pitch contrast in both speech stimuli (syllable /ji/ in Experiment 1) and nonspeech stimuli (harmonic tone in Experiment 1; pure tone in Experiment 2) was presented to native Cantonese speakers in passive oddball paradigms. We found that the mismatch negativity (MMN) elicited by lexical pitch contrast was lateralized to the left hemisphere, which is consistent with the pattern of function-dependent brain asymmetry (i.e., left hemisphere lateralization for speech processing) in nontonal language speakers. However, the MMN elicited by acoustic pitch contrast was also left hemisphere lateralized (harmonic tone in Experiment 1) or showed a tendency for left hemisphere lateralization (pure tone in Experiment 2), which is inconsistent with the pattern of acoustic-dependent brain asymmetry (i.e., right hemisphere lateralization for acoustic pitch processing) in nontonal language speakers. The consistent pattern of function-dependent brain asymmetry and the inconsistent pattern of acoustic-dependent brain asymmetry between tonal and nontonal language speakers can be explained by the hypothesis that the acoustic-dependent brain asymmetry is the consequence of a carryover effect from function-dependent brain asymmetry. Potential evolutionary implication of this hypothesis is discussed. © 2013.

  2. Compound nouns in spoken language production by speakers with aphasia compared to neurologically healthy speakers: an exploratory study.

    PubMed

    Eiesland, Eli Anne; Lind, Marianne

    2012-03-01

    Compounds are words that are made up of at least two other words (lexemes), featuring lexical and syntactic characteristics and thus particularly interesting for the study of language processing. Most studies of compounds and language processing have been based on data from experimental single word production and comprehension tasks. To enhance the ecological validity of morphological processing research, data from other contexts, such as discourse production, need to be considered. This study investigates the production of nominal compounds in semi-spontaneous spoken texts by a group of speakers with fluent types of aphasia compared to a group of neurologically healthy speakers. The speakers with aphasia produce significantly fewer nominal compound types in their texts than the non-aphasic speakers, and the compounds they produce exhibit fewer different types of semantic relations than the compounds produced by the non-aphasic speakers. The results are discussed in relation to theories of language processing.

  3. Implementation of support vector machine for classification of speech marked hijaiyah letters based on Mel frequency cepstrum coefficient feature extraction

    NASA Astrophysics Data System (ADS)

    Adhi Pradana, Wisnu; Adiwijaya; Novia Wisesty, Untari

    2018-03-01

    Support Vector Machine or commonly called SVM is one method that can be used to process the classification of a data. SVM classifies data from 2 different classes with hyperplane. In this study, the system was built using SVM to develop Arabic Speech Recognition. In the development of the system, there are 2 kinds of speakers that have been tested that is dependent speakers and independent speakers. The results from this system is an accuracy of 85.32% for speaker dependent and 61.16% for independent speakers.

  4. Statistical Evaluation of Biometric Evidence in Forensic Automatic Speaker Recognition

    NASA Astrophysics Data System (ADS)

    Drygajlo, Andrzej

    Forensic speaker recognition is the process of determining if a specific individual (suspected speaker) is the source of a questioned voice recording (trace). This paper aims at presenting forensic automatic speaker recognition (FASR) methods that provide a coherent way of quantifying and presenting recorded voice as biometric evidence. In such methods, the biometric evidence consists of the quantified degree of similarity between speaker-dependent features extracted from the trace and speaker-dependent features extracted from recorded speech of a suspect. The interpretation of recorded voice as evidence in the forensic context presents particular challenges, including within-speaker (within-source) variability and between-speakers (between-sources) variability. Consequently, FASR methods must provide a statistical evaluation which gives the court an indication of the strength of the evidence given the estimated within-source and between-sources variabilities. This paper reports on the first ENFSI evaluation campaign through a fake case, organized by the Netherlands Forensic Institute (NFI), as an example, where an automatic method using the Gaussian mixture models (GMMs) and the Bayesian interpretation (BI) framework were implemented for the forensic speaker recognition task.

  5. Semi-spontaneous oral text production: measurements in clinical practice.

    PubMed

    Lind, Marianne; Kristoffersen, Kristian Emil; Moen, Inger; Simonsen, Hanne Gram

    2009-12-01

    Functionally relevant assessment of the language production of speakers with aphasia should include assessment of connected speech production. Despite the ecological validity of everyday conversations, more controlled and monological types of texts may be easier to obtain and analyse in clinical practice. This article discusses some simple measurements for the analysis of semi-spontaneous oral text production by speakers with aphasia. Specifically, the measurements are related to the production of verbs and nouns, and the realization of different sentence types. The proposed measurements should be clinically relevant, easily applicable, and linguistically meaningful. The measurements have been applied to oral descriptions of the 'Cookie Theft' picture by eight monolingual Norwegian speakers, four with an anomic type of aphasia and four without any type of language impairment. Despite individual differences in both the clinical and the non-clinical group, most of the measurements seem to distinguish between speakers with and without aphasia.

  6. Syntactic Complexity and L2 Academic Immersion Effects on Readers' Recall and Pausing Strategies for English and Spanish Texts

    ERIC Educational Resources Information Center

    Lazarte, Alejandro A.; Barry, Sue

    2008-01-01

    In Experiment 1, monolingual native Spanish speakers (NSSs) had better kernel recall and longer end-of-clause (EOC) pauses than native English speakers (NESs) when reading texts that varied in syntactic complexity as a function of the number of nonessential clauses added to the kernel text. NSS familiarity with embedded clauses in Spanish seem to…

  7. The 2016 NIST Speaker Recognition Evaluation

    DTIC Science & Technology

    2017-08-20

    The 2016 NIST Speaker Recognition Evaluation Seyed Omid Sadjadi1,∗, Timothée Kheyrkhah1,†, Audrey Tong1, Craig Greenberg1, Douglas Reynolds2, Elliot...recent in an ongoing series of speaker recognition evaluations (SRE) to foster research in ro- bust text-independent speaker recognition, as well as...online evaluation platform, a fixed training data condition, more variability in test segment duration (uni- formly distributed between 10s and 60s

  8. Promoting Communities of Practice among Non-Native Speakers of English in Online Discussions

    ERIC Educational Resources Information Center

    Kim, Hoe Kyeung

    2011-01-01

    An online discussion involving text-based computer-mediated communication has great potential for promoting equal participation among non-native speakers of English. Several studies claimed that online discussions could enhance the academic participation of non-native speakers of English. However, there is little research around participation…

  9. ASTP Technical Air-To-Ground Voice Transcription

    NASA Technical Reports Server (NTRS)

    1975-01-01

    The transcription of the technical air-to-ground voice communication of the Apollo-Soyuz Test Project mission was presented. The transcript was divided into three columns giving, respectively, the time, speaker, and text. All times are expressed in Greenwich mean time for the appropriate Julian dates. The speaker column indicates the source of transmission; the text column contains the verbatim transcript of the communications. Special symbols were used to report garbling, pauses or self-interruptions, interruptions by other speakers or abrupt terminations, emphasized words, obliterations, and material translated from Russian.

  10. What's Learned Together Stays Together: Speakers' Choice of Referring Expression Reflects Shared Experience

    ERIC Educational Resources Information Center

    Gorman, Kristen S.; Gegg-Harrison, Whitney; Marsh, Chelsea R.; Tanenhaus, Michael K.

    2013-01-01

    When referring to named objects, speakers can choose either a name ("mbira") or a description ("that gourd-like instrument with metal strips"); whether the name provides useful information depends on whether the speaker's knowledge of the name is shared with the addressee. But, how do speakers determine what is shared? In 2…

  11. Recognition of speaker-dependent continuous speech with KEAL

    NASA Astrophysics Data System (ADS)

    Mercier, G.; Bigorgne, D.; Miclet, L.; Le Guennec, L.; Querre, M.

    1989-04-01

    A description of the speaker-dependent continuous speech recognition system KEAL is given. An unknown utterance, is recognized by means of the followng procedures: acoustic analysis, phonetic segmentation and identification, word and sentence analysis. The combination of feature-based, speaker-independent coarse phonetic segmentation with speaker-dependent statistical classification techniques is one of the main design features of the acoustic-phonetic decoder. The lexical access component is essentially based on a statistical dynamic programming technique which aims at matching a phonemic lexical entry containing various phonological forms, against a phonetic lattice. Sentence recognition is achieved by use of a context-free grammar and a parsing algorithm derived from Earley's parser. A speaker adaptation module allows some of the system parameters to be adjusted by matching known utterances with their acoustical representation. The task to be performed, described by its vocabulary and its grammar, is given as a parameter of the system. Continuously spoken sentences extracted from a 'pseudo-Logo' language are analyzed and results are presented.

  12. Speaker normalization for chinese vowel recognition in cochlear implants.

    PubMed

    Luo, Xin; Fu, Qian-Jie

    2005-07-01

    Because of the limited spectra-temporal resolution associated with cochlear implants, implant patients often have greater difficulty with multitalker speech recognition. The present study investigated whether multitalker speech recognition can be improved by applying speaker normalization techniques to cochlear implant speech processing. Multitalker Chinese vowel recognition was tested with normal-hearing Chinese-speaking subjects listening to a 4-channel cochlear implant simulation, with and without speaker normalization. For each subject, speaker normalization was referenced to the speaker that produced the best recognition performance under conditions without speaker normalization. To match the remaining speakers to this "optimal" output pattern, the overall frequency range of the analysis filter bank was adjusted for each speaker according to the ratio of the mean third formant frequency values between the specific speaker and the reference speaker. Results showed that speaker normalization provided a small but significant improvement in subjects' overall recognition performance. After speaker normalization, subjects' patterns of recognition performance across speakers changed, demonstrating the potential for speaker-dependent effects with the proposed normalization technique.

  13. The Sound of Voice: Voice-Based Categorization of Speakers' Sexual Orientation within and across Languages.

    PubMed

    Sulpizio, Simone; Fasoli, Fabio; Maass, Anne; Paladino, Maria Paola; Vespignani, Francesco; Eyssel, Friederike; Bentler, Dominik

    2015-01-01

    Empirical research had initially shown that English listeners are able to identify the speakers' sexual orientation based on voice cues alone. However, the accuracy of this voice-based categorization, as well as its generalizability to other languages (language-dependency) and to non-native speakers (language-specificity), has been questioned recently. Consequently, we address these open issues in 5 experiments: First, we tested whether Italian and German listeners are able to correctly identify sexual orientation of same-language male speakers. Then, participants of both nationalities listened to voice samples and rated the sexual orientation of both Italian and German male speakers. We found that listeners were unable to identify the speakers' sexual orientation correctly. However, speakers were consistently categorized as either heterosexual or gay on the basis of how they sounded. Moreover, a similar pattern of results emerged when listeners judged the sexual orientation of speakers of their own and of the foreign language. Overall, this research suggests that voice-based categorization of sexual orientation reflects the listeners' expectations of how gay voices sound rather than being an accurate detector of the speakers' actual sexual identity. Results are discussed with regard to accuracy, acoustic features of voices, language dependency and language specificity.

  14. Neural Systems Involved When Attending to a Speaker

    PubMed Central

    Kamourieh, Salwa; Braga, Rodrigo M.; Leech, Robert; Newbould, Rexford D.; Malhotra, Paresh; Wise, Richard J. S.

    2015-01-01

    Remembering what a speaker said depends on attention. During conversational speech, the emphasis is on working memory, but listening to a lecture encourages episodic memory encoding. With simultaneous interference from background speech, the need for auditory vigilance increases. We recreated these context-dependent demands on auditory attention in 2 ways. The first was to require participants to attend to one speaker in either the absence or presence of a distracting background speaker. The second was to alter the task demand, requiring either an immediate or delayed recall of the content of the attended speech. Across 2 fMRI studies, common activated regions associated with segregating attended from unattended speech were the right anterior insula and adjacent frontal operculum (aI/FOp), the left planum temporale, and the precuneus. In contrast, activity in a ventral right frontoparietal system was dependent on both the task demand and the presence of a competing speaker. Additional multivariate analyses identified other domain-general frontoparietal systems, where activity increased during attentive listening but was modulated little by the need for speech stream segregation in the presence of 2 speakers. These results make predictions about impairments in attentive listening in different communicative contexts following focal or diffuse brain pathology. PMID:25596592

  15. Continuing Medical Education Speakers with High Evaluation Scores Use more Image-based Slides.

    PubMed

    Ferguson, Ian; Phillips, Andrew W; Lin, Michelle

    2017-01-01

    Although continuing medical education (CME) presentations are common across health professions, it is unknown whether slide design is independently associated with audience evaluations of the speaker. Based on the conceptual framework of Mayer's theory of multimedia learning, this study aimed to determine whether image use and text density in presentation slides are associated with overall speaker evaluations. This retrospective analysis of six sequential CME conferences (two annual emergency medicine conferences over a three-year period) used a mixed linear regression model to assess whether post-conference speaker evaluations were associated with image fraction (percentage of image-based slides per presentation) and text density (number of words per slide). A total of 105 unique lectures were given by 49 faculty members, and 1,222 evaluations (70.1% response rate) were available for analysis. On average, 47.4% (SD=25.36) of slides had at least one educationally-relevant image (image fraction). Image fraction significantly predicted overall higher evaluation scores [F(1, 100.676)=6.158, p=0.015] in the mixed linear regression model. The mean (SD) text density was 25.61 (8.14) words/slide but was not a significant predictor [F(1, 86.293)=0.55, p=0.815]. Of note, the individual speaker [χ 2 (1)=2.952, p=0.003] and speaker seniority [F(3, 59.713)=4.083, p=0.011] significantly predicted higher scores. This is the first published study to date assessing the linkage between slide design and CME speaker evaluations by an audience of practicing clinicians. The incorporation of images was associated with higher evaluation scores, in alignment with Mayer's theory of multimedia learning. Contrary to this theory, however, text density showed no significant association, suggesting that these scores may be multifactorial. Professional development efforts should focus on teaching best practices in both slide design and presentation skills.

  16. On the optimization of a mixed speaker array in an enclosed space using the virtual-speaker weighting method

    NASA Astrophysics Data System (ADS)

    Peng, Bo; Zheng, Sifa; Liao, Xiangning; Lian, Xiaomin

    2018-03-01

    In order to achieve sound field reproduction in a wide frequency band, multiple-type speakers are used. The reproduction accuracy is not only affected by the signals sent to the speakers, but also depends on the position and the number of each type of speaker. The method of optimizing a mixed speaker array is investigated in this paper. A virtual-speaker weighting method is proposed to optimize both the position and the number of each type of speaker. In this method, a virtual-speaker model is proposed to quantify the increment of controllability of the speaker array when the speaker number increases. While optimizing a mixed speaker array, the gain of the virtual-speaker transfer function is used to determine the priority orders of the candidate speaker positions, which optimizes the position of each type of speaker. Then the relative gain of the virtual-speaker transfer function is used to determine whether the speakers are redundant, which optimizes the number of each type of speaker. Finally the virtual-speaker weighting method is verified by reproduction experiments of the interior sound field in a passenger car. The results validate that the optimum mixed speaker array can be obtained using the proposed method.

  17. Tier-Adjacency Is Not a Necessary Condition for Learning Phonotactic Dependencies

    ERIC Educational Resources Information Center

    Koo, Hahn; Callahan, Lydia

    2012-01-01

    One hypothesis raised by Newport and Aslin to explain how speakers learn dependencies between nonadjacent phonemes is that speakers track bigram probabilities between two segments that are adjacent to each other within a tier of their own. The hypothesis predicts that a dependency between segments separated from each other at the tier level cannot…

  18. Persian Native Speakers Reading Persian and English Texts: Their Strategic Behavior to Overcome Syntactic and Semantic Problems

    ERIC Educational Resources Information Center

    Alimorad, Zahra

    2015-01-01

    This study aimed to discover semantic and syntactic problems Persian native speakers might have while reading English and Persian texts and different strategies they use to overcome those problems. To this end, a convenient sample of 40 intermediate students studying English Literature at Shiraz University was selected. Twenty of them were asked…

  19. The Sound of Voice: Voice-Based Categorization of Speakers’ Sexual Orientation within and across Languages

    PubMed Central

    Maass, Anne; Paladino, Maria Paola; Vespignani, Francesco; Eyssel, Friederike; Bentler, Dominik

    2015-01-01

    Empirical research had initially shown that English listeners are able to identify the speakers' sexual orientation based on voice cues alone. However, the accuracy of this voice-based categorization, as well as its generalizability to other languages (language-dependency) and to non-native speakers (language-specificity), has been questioned recently. Consequently, we address these open issues in 5 experiments: First, we tested whether Italian and German listeners are able to correctly identify sexual orientation of same-language male speakers. Then, participants of both nationalities listened to voice samples and rated the sexual orientation of both Italian and German male speakers. We found that listeners were unable to identify the speakers' sexual orientation correctly. However, speakers were consistently categorized as either heterosexual or gay on the basis of how they sounded. Moreover, a similar pattern of results emerged when listeners judged the sexual orientation of speakers of their own and of the foreign language. Overall, this research suggests that voice-based categorization of sexual orientation reflects the listeners' expectations of how gay voices sound rather than being an accurate detector of the speakers' actual sexual identity. Results are discussed with regard to accuracy, acoustic features of voices, language dependency and language specificity. PMID:26132820

  20. Speaker identification for the improvement of the security communication between law enforcement units

    NASA Astrophysics Data System (ADS)

    Tovarek, Jaromir; Partila, Pavol

    2017-05-01

    This article discusses the speaker identification for the improvement of the security communication between law enforcement units. The main task of this research was to develop the text-independent speaker identification system which can be used for real-time recognition. This system is designed for identification in the open set. It means that the unknown speaker can be anyone. Communication itself is secured, but we have to check the authorization of the communication parties. We have to decide if the unknown speaker is the authorized for the given action. The calls are recorded by IP telephony server and then these recordings are evaluate using classification If the system evaluates that the speaker is not authorized, it sends a warning message to the administrator. This message can detect, for example a stolen phone or other unusual situation. The administrator then performs the appropriate actions. Our novel proposal system uses multilayer neural network for classification and it consists of three layers (input layer, hidden layer, and output layer). A number of neurons in input layer corresponds with the length of speech features. Output layer then represents classified speakers. Artificial Neural Network classifies speech signal frame by frame, but the final decision is done over the complete record. This rule substantially increases accuracy of the classification. Input data for the neural network are a thirteen Mel-frequency cepstral coefficients, which describe the behavior of the vocal tract. These parameters are the most used for speaker recognition. Parameters for training, testing and validation were extracted from recordings of authorized users. Recording conditions for training data correspond with the real traffic of the system (sampling frequency, bit rate). The main benefit of the research is the system developed for text-independent speaker identification which is applied to secure communication between law enforcement units.

  1. Analysis of human scream and its impact on text-independent speaker verification.

    PubMed

    Hansen, John H L; Nandwana, Mahesh Kumar; Shokouhi, Navid

    2017-04-01

    Scream is defined as sustained, high-energy vocalizations that lack phonological structure. Lack of phonological structure is how scream is identified from other forms of loud vocalization, such as "yell." This study investigates the acoustic aspects of screams and addresses those that are known to prevent standard speaker identification systems from recognizing the identity of screaming speakers. It is well established that speaker variability due to changes in vocal effort and Lombard effect contribute to degraded performance in automatic speech systems (i.e., speech recognition, speaker identification, diarization, etc.). However, previous research in the general area of speaker variability has concentrated on human speech production, whereas less is known about non-speech vocalizations. The UT-NonSpeech corpus is developed here to investigate speaker verification from scream samples. This study considers a detailed analysis in terms of fundamental frequency, spectral peak shift, frame energy distribution, and spectral tilt. It is shown that traditional speaker recognition based on the Gaussian mixture models-universal background model framework is unreliable when evaluated with screams.

  2. Native and Non-Native Speakers' Brain Responses to Filled Indirect Object Gaps

    ERIC Educational Resources Information Center

    Jessen, Anna; Festman, Julia; Boxell, Oliver; Felser, Claudia

    2017-01-01

    We examined native and non-native English speakers' processing of indirect object "wh"-dependencies using a filled-gap paradigm while recording event-related potentials (ERPs). The non-native group was comprised of native German-speaking, proficient non-native speakers of English. Both participant groups showed evidence of linking…

  3. Speech variability effects on recognition accuracy associated with concurrent task performance by pilots

    NASA Technical Reports Server (NTRS)

    Simpson, C. A.

    1985-01-01

    In the present study of the responses of pairs of pilots to aircraft warning classification tasks using an isolated word, speaker-dependent speech recognition system, the induced stress was manipulated by means of different scoring procedures for the classification task and by the inclusion of a competitive manual control task. Both speech patterns and recognition accuracy were analyzed, and recognition errors were recorded by type for an isolated word speaker-dependent system and by an offline technique for a connected word speaker-dependent system. While errors increased with task loading for the isolated word system, there was no such effect for task loading in the case of the connected word system.

  4. The role of linguistic experience in the processing of probabilistic information in production.

    PubMed

    Gustafson, Erin; Goldrick, Matthew

    2018-01-01

    Speakers track the probability that a word will occur in a particular context and utilize this information during phonetic processing. For example, content words that have high probability within a discourse tend to be realized with reduced acoustic/articulatory properties. Such probabilistic information may influence L1 and L2 speech processing in distinct ways (reflecting differences in linguistic experience across groups and the overall difficulty of L2 speech processing). To examine this issue, L1 and L2 speakers performed a referential communication task, describing sequences of simple actions. The two groups of speakers showed similar effects of discourse-dependent probabilistic information on production, suggesting that L2 speakers can successfully track discourse-dependent probabilities and use such information to modulate phonetic processing.

  5. And then I saw her race: Race-based expectations affect infants' word processing.

    PubMed

    Weatherhead, Drew; White, Katherine S

    2018-08-01

    How do our expectations about speakers shape speech perception? Adults' speech perception is influenced by social properties of the speaker (e.g., race). When in development do these influences begin? In the current study, 16-month-olds heard familiar words produced in their native accent (e.g., "dog") and in an unfamiliar accent involving a vowel shift (e.g., "dag"), in the context of an image of either a same-race speaker or an other-race speaker. Infants' interpretation of the words depended on the speaker's race. For the same-race speaker, infants only recognized words produced in the familiar accent; for the other-race speaker, infants recognized both versions of the words. Two additional experiments showed that infants only recognized an other-race speaker's atypical pronunciations when they differed systematically from the native accent. These results provide the first evidence that expectations driven by unspoken properties of speakers, such as race, influence infants' speech processing. Copyright © 2018 Elsevier B.V. All rights reserved.

  6. Do Listeners Store in Memory a Speaker's Habitual Utterance-Final Phonation Type?

    PubMed Central

    Bőhm, Tamás; Shattuck-Hufnagel, Stefanie

    2009-01-01

    Earlier studies report systematic differences across speakers in the occurrence of utterance-final irregular phonation; the work reported here investigated whether human listeners remember this speaker-specific information and can access it when necessary (a prerequisite for using this cue in speaker recognition). Listeners personally familiar with the voices of the speakers were presented with pairs of speech samples: one with the original and the other with transformed final phonation type. Asked to select the member of the pair that was closer to the talker's voice, most listeners tended to choose the unmanipulated token (even though they judged them to sound essentially equally natural). This suggests that utterance-final pitch period irregularity is part of the mental representation of individual speaker voices, although this may depend on the individual speaker and listener to some extent. PMID:19776665

  7. Native and Nonnative Speakers' Pragmatic Interpretations of English Texts.

    ERIC Educational Resources Information Center

    Hinkel, Eli

    1994-01-01

    Considering the complicating effect of cultural differences in writing conventions, this study examines discourse tradition as influenced by Confucian/Taoist precepts and those of U.S. academic environments, the latter requiring rational argumentation, justification, and proof. Pedagogical implications of native-speaker and nonnative-speaker…

  8. Unlocking Academic Vocabulary

    ERIC Educational Resources Information Center

    DeLuca, Eileen

    2010-01-01

    How can we teach science to English language learners (ELLs) when even our native English speakers have trouble reading the textbook? To help science teachers meet this challenge, this article presents six text-comprehension strategies used by English for Speakers of other Languages (ESOL) teachers: metalinguistic awareness development,…

  9. Syntactic Constraints and Individual Differences in Native and Non-Native Processing of Wh-Movement

    PubMed Central

    Johnson, Adrienne; Fiorentino, Robert; Gabriele, Alison

    2016-01-01

    There is a debate as to whether second language (L2) learners show qualitatively similar processing profiles as native speakers or whether L2 learners are restricted in their ability to use syntactic information during online processing. In the realm of wh-dependency resolution, research has examined whether learners, similar to native speakers, attempt to resolve wh-dependencies in grammatically licensed contexts but avoid positing gaps in illicit contexts such as islands. Also at issue is whether the avoidance of gap filling in islands is due to adherence to syntactic constraints or whether islands simply present processing bottlenecks. One approach has been to examine the relationship between processing abilities and the establishment of wh-dependencies in islands. Grammatical accounts of islands do not predict such a relationship as the parser should simply not predict gaps in illicit contexts. In contrast, a pattern of results showing that individuals with more processing resources are better able to establish wh-dependencies in islands could conceivably be compatible with certain processing accounts. In a self-paced reading experiment which examines the processing of wh-dependencies, we address both questions, examining whether native English speakers and Korean learners of English show qualitatively similar patterns and whether there is a relationship between working memory, as measured by counting span and reading span, and processing in both island and non-island contexts. The results of the self-paced reading experiment suggest that learners can use syntactic information on the same timecourse as native speakers, showing qualitative similarity between the two groups. Results of regression analyses did not reveal a significant relationship between working memory and the establishment of wh-dependencies in islands but we did observe significant relationships between working memory and the processing of licit wh-dependencies. As the contexts in which these relationships emerged differed for learners and native speakers, our results call for further research examining individual differences in dependency resolution in both populations. PMID:27148152

  10. Syntactic Constraints and Individual Differences in Native and Non-Native Processing of Wh-Movement.

    PubMed

    Johnson, Adrienne; Fiorentino, Robert; Gabriele, Alison

    2016-01-01

    There is a debate as to whether second language (L2) learners show qualitatively similar processing profiles as native speakers or whether L2 learners are restricted in their ability to use syntactic information during online processing. In the realm of wh-dependency resolution, research has examined whether learners, similar to native speakers, attempt to resolve wh-dependencies in grammatically licensed contexts but avoid positing gaps in illicit contexts such as islands. Also at issue is whether the avoidance of gap filling in islands is due to adherence to syntactic constraints or whether islands simply present processing bottlenecks. One approach has been to examine the relationship between processing abilities and the establishment of wh-dependencies in islands. Grammatical accounts of islands do not predict such a relationship as the parser should simply not predict gaps in illicit contexts. In contrast, a pattern of results showing that individuals with more processing resources are better able to establish wh-dependencies in islands could conceivably be compatible with certain processing accounts. In a self-paced reading experiment which examines the processing of wh-dependencies, we address both questions, examining whether native English speakers and Korean learners of English show qualitatively similar patterns and whether there is a relationship between working memory, as measured by counting span and reading span, and processing in both island and non-island contexts. The results of the self-paced reading experiment suggest that learners can use syntactic information on the same timecourse as native speakers, showing qualitative similarity between the two groups. Results of regression analyses did not reveal a significant relationship between working memory and the establishment of wh-dependencies in islands but we did observe significant relationships between working memory and the processing of licit wh-dependencies. As the contexts in which these relationships emerged differed for learners and native speakers, our results call for further research examining individual differences in dependency resolution in both populations.

  11. Building Searchable Collections of Enterprise Speech Data.

    ERIC Educational Resources Information Center

    Cooper, James W.; Viswanathan, Mahesh; Byron, Donna; Chan, Margaret

    The study has applied speech recognition and text-mining technologies to a set of recorded outbound marketing calls and analyzed the results. Since speaker-independent speech recognition technology results in a significantly lower recognition rate than that found when the recognizer is trained for a particular speaker, a number of post-processing…

  12. Voice Recognition Software Accuracy with Second Language Speakers of English.

    ERIC Educational Resources Information Center

    Coniam, D.

    1999-01-01

    Explores the potential of the use of voice-recognition technology with second-language speakers of English. Involves the analysis of the output produced by a small group of very competent second-language subjects reading a text into the voice recognition software Dragon Systems "Dragon NaturallySpeaking." (Author/VWL)

  13. Developing Communication in the Workplace for Non-Native English Speakers.

    ERIC Educational Resources Information Center

    Nichols, Pat; Watkins, Lisa

    This curriculum module contains materials for conducting a course designed to build oral and written English skills for nonnative speakers. The course focuses on increasing vocabulary, improving listening/speaking skills, extracting information from various written texts (such as memos, notes, business forms, manuals, letters), and developing…

  14. Qualitative Feedback From a Text Messaging Intervention for Depression: Benefits, Drawbacks, and Cultural Differences

    PubMed Central

    Berridge, Clara

    2014-01-01

    Background Mobile health interventions are often standardized and assumed to work the same for all users; however, we may be missing cultural differences in the experiences of interventions that may impact how and if an intervention is effective. Objective The objective of the study was to assess qualitative feedback from participants to determine if there were differences between Spanish speakers and English speakers. Daily text messages were sent to patients as an adjunct to group Cognitive Behavioral Therapy (CBT) for depression. Methods Messages inquired about mood and about specific themes (thoughts, activities, social interactions) of a manualized group CBT intervention. There were thirty-nine patients who participated in the text messaging pilot study. The average age of the participants was 53 years (SD 10.4; range of 23-72). Results Qualitative feedback from Spanish speakers highlighted feelings of social support, whereas English speakers noted increased introspection and self-awareness of their mood state. Conclusions These cultural differences should be explored further, as they may impact the effect of supportive mobile health interventions. Trial Registration Trial Registration: Clinicaltrials.gov NCT01083628; http://clinicaltrials.gov/ct2/show/study/NCT01083628 (Archived by WebCite at http://www.webcitation.org/6StpbdHuq). PMID:25373390

  15. 75 FR 61440 - 36(b)(1) Arms Sales Notifications

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-10-05

    ... Department of Defense is publishing the unclassified text of three section 36(b)(1) arms sales notifications... are copies of letters to the Speaker of the House of Representatives, Transmittals 10-20, 10-23, and... a letter to the Speaker of the House of Representatives, Transmittal 10-47 with attached transmittal...

  16. 75 FR 20571 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-04-20

    ... Department of Defense is publishing the unclassified text of two section 36(b)(1) arms sales notifications to... of letters to the Speaker of the House of Representatives, Transmittals 10-04 and 10-14 with... copy of a letter to the Speaker of the House of Representatives, Transmittals 10-04 with attached...

  17. 75 FR 41820 - 36(b)(1) Arms Sales Notifications

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-07-19

    ...: Notice. SUMMARY: The Department of Defense is publishing the unclassified text of five section 36(b)(1... INFORMATION: The following are copies of letters to the Speaker of the House of Representatives, Transmittals... following is a copy of a letter to the Speaker of the House of Representatives, Transmittal 10-05 with...

  18. Writing for the Ear: Strengthening Oral Style in Manuscript Speeches

    ERIC Educational Resources Information Center

    Bruss, Kristine

    2012-01-01

    Public speaking texts typically advise speakers to avoid using a manuscript. Speaking from a manuscript can limit eye contact, reduce expressiveness, and bore listeners. The ideal, rather, is to sound conversational. Conversational style is inclusive, suggesting that a speaker is ""of the people," united in understanding, values and purpose." If a…

  19. Rhetoric and the Interpretation of Poetry.

    ERIC Educational Resources Information Center

    Sloan, Thomas O.

    The ability to think of poetry rhetorically is a valuable instrument for interpreting poetry. The poet is the speaker "of" the poem, the persona the speaker "in" the poem. The communicative circle is complete when it includes the reader who combines an analysis of the text (the words of the persona) with an analysis of the…

  20. Partially supervised speaker clustering.

    PubMed

    Tang, Hao; Chu, Stephen Mingyu; Hasegawa-Johnson, Mark; Huang, Thomas S

    2012-05-01

    Content-based multimedia indexing, retrieval, and processing as well as multimedia databases demand the structuring of the media content (image, audio, video, text, etc.), one significant goal being to associate the identity of the content to the individual segments of the signals. In this paper, we specifically address the problem of speaker clustering, the task of assigning every speech utterance in an audio stream to its speaker. We offer a complete treatment to the idea of partially supervised speaker clustering, which refers to the use of our prior knowledge of speakers in general to assist the unsupervised speaker clustering process. By means of an independent training data set, we encode the prior knowledge at the various stages of the speaker clustering pipeline via 1) learning a speaker-discriminative acoustic feature transformation, 2) learning a universal speaker prior model, and 3) learning a discriminative speaker subspace, or equivalently, a speaker-discriminative distance metric. We study the directional scattering property of the Gaussian mixture model (GMM) mean supervector representation of utterances in the high-dimensional space, and advocate exploiting this property by using the cosine distance metric instead of the euclidean distance metric for speaker clustering in the GMM mean supervector space. We propose to perform discriminant analysis based on the cosine distance metric, which leads to a novel distance metric learning algorithm—linear spherical discriminant analysis (LSDA). We show that the proposed LSDA formulation can be systematically solved within the elegant graph embedding general dimensionality reduction framework. Our speaker clustering experiments on the GALE database clearly indicate that 1) our speaker clustering methods based on the GMM mean supervector representation and vector-based distance metrics outperform traditional speaker clustering methods based on the “bag of acoustic features” representation and statistical model-based distance metrics, 2) our advocated use of the cosine distance metric yields consistent increases in the speaker clustering performance as compared to the commonly used euclidean distance metric, 3) our partially supervised speaker clustering concept and strategies significantly improve the speaker clustering performance over the baselines, and 4) our proposed LSDA algorithm further leads to state-of-the-art speaker clustering performance.

  1. Multimedia proceedings of the 10th Office Information Technology Conference

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hudson, B.

    1993-09-10

    The CD contains the handouts for all the speakers, demo software from Apple, Adobe, Microsoft, and Zylabs, and video movies of the keynote speakers. Adobe Acrobat is used to provide full-fidelity retrieval of the speakers` slides and Apple`s Quicktime for Macintosh and Windows is used for video playback. ZyIndex is included for Windows users to provide a full-text search engine for selected documents. There are separately labelled installation and operating instructions for Macintosh and Windows users and some general materials common to both sets of users.

  2. Social dominance orientation, nonnative accents, and hiring recommendations.

    PubMed

    Hansen, Karolina; Dovidio, John F

    2016-10-01

    Discrimination against nonnative speakers is widespread and largely socially acceptable. Nonnative speakers are evaluated negatively because accent is a sign that they belong to an outgroup and because understanding their speech requires unusual effort from listeners. The present research investigated intergroup bias, based on stronger support for hierarchical relations between groups (social dominance orientation [SDO]), as a predictor of hiring recommendations of nonnative speakers. In an online experiment using an adaptation of the thin-slices methodology, 65 U.S. adults (54% women; 80% White; Mage = 35.91, range = 18-67) heard a recording of a job applicant speaking with an Asian (Mandarin Chinese) or a Latino (Spanish) accent. Participants indicated how likely they would be to recommend hiring the speaker, answered questions about the text, and indicated how difficult it was to understand the applicant. Independent of objective comprehension, participants high in SDO reported that it was more difficult to understand a Latino speaker than an Asian speaker. SDO predicted hiring recommendations of the speakers, but this relationship was mediated by the perception that nonnative speakers were difficult to understand. This effect was stronger for speakers from lower status groups (Latinos relative to Asians) and was not related to objective comprehension. These findings suggest a cycle of prejudice toward nonnative speakers: Not only do perceptions of difficulty in understanding cause prejudice toward them, but also prejudice toward low-status groups can lead to perceived difficulty in understanding members of these groups. (PsycINFO Database Record (c) 2016 APA, all rights reserved).

  3. A Computer Text Analysis of Four Cohesion Devices in English Discourse by Native and Nonnative Writers.

    ERIC Educational Resources Information Center

    Reid, Joy

    1992-01-01

    In a contrastive rhetoric study of nonnative English speakers, 768 essays written in English by native speakers of Arabic, Chinese, Spanish, and English were examined using the Writer's Workbench program to determine whether distinctive, quantifiable differences in the use of 4 cohesion devices existed among the 4 language backgrounds. (Author/LB)

  4. Effects of Instruction on Chinese College Students' Thematic Choice in Academic Writing

    ERIC Educational Resources Information Center

    Wei, Jing

    2016-01-01

    The Theme is a major aspect of how speakers construct their messages in a way which makes them fit smoothly into the unfolding language event. Thematic choice provides clues as to how English learners organize information and shape their texts. Previous studies reveal that English learners deviated from English native speakers in their thematic…

  5. Apprendre l'orthographe avec un correcteur orthographique (Learning Spelling with a Spell-Checker?)?

    ERIC Educational Resources Information Center

    Desmarais, Lise

    1998-01-01

    Reports a study with 27 adults, both native French-speakers and native English-speakers, on the effectiveness of using a spell-checker as the core element to teach French spelling. The method used authentic materials, individualized monitoring, screen and hard-copy text reading, and content sequencing based on errors. The approach generated…

  6. Sinteiseoir 1.0: A Multidialectical TTS Application for Irish

    ERIC Educational Resources Information Center

    Mac Lochlainn, Micheal

    2010-01-01

    This paper details the development of a multidialectical text-to-speech (TTS) application, "Sinteiseoir," for the Irish language. This work is being carried out in the context of Irish as a lesser-used language, where learners and other L2 speakers have limited direct exposure to L1 speakers and speech communities, and where native sound…

  7. "My Major Is English, Believe It or Not:)" -- Participant Orientations in Nonnative/Native Text Chat

    ERIC Educational Resources Information Center

    Vandergriff, Ilona

    2013-01-01

    In their interactions with native speakers (NS), nonnative speakers (NNS) often position themselves as relative novices. For example, they may orient to the language expertise differential by apologizing for their linguistic ineptness or by making self-disparaging remarks about their second language (L2). This is true even for advanced learners in…

  8. Processing of Written Irony in Autism Spectrum Disorder: An Eye-Movement Study.

    PubMed

    Au-Yeung, Sheena K; Kaakinen, Johanna K; Liversedge, Simon P; Benson, Valerie

    2015-12-01

    Previous research has suggested that individuals with Autism Spectrum Disorders (ASD) have difficulties understanding others communicative intent and with using contextual information to correctly interpret irony. We recorded the eye movements of typically developing (TD) adults ASD adults when they read statements that could either be interpreted as ironic or non-ironic depending on the context of the passage. Participants with ASD performed as well as TD controls in their comprehension accuracy for speaker's statements in both ironic and non-ironic conditions. Eye movement data showed that for both participant groups, total reading times were longer for the critical region containing the speaker's statement and a subsequent sentence restating the context in the ironic condition compared to the non-ironic condition. The results suggest that more effortful processing is required in both ASD and TD participants for ironic compared with literal non-ironic statements, and that individuals with ASD were able to use contextual information to infer a non-literal interpretation of ironic text. Individuals with ASD, however, spent more time overall than TD controls rereading the passages, to a similar degree across both ironic and non-ironic conditions, suggesting that they either take longer to construct a coherent discourse representation of the text, or that they take longer to make the decision that their representation of the text is reasonable based on their knowledge of the world. © 2015 International Society for Autism Research, Wiley Periodicals, Inc.

  9. What a speaker's choice of frame reveals: reference points, frame selection, and framing effects.

    PubMed

    McKenzie, Craig R M; Nelson, Jonathan D

    2003-09-01

    Framing effects are well established: Listeners' preferences depend on how outcomes are described to them, or framed. Less well understood is what determines how speakers choose frames. Two experiments revealed that reference points systematically influenced speakers' choices between logically equivalent frames. For example, speakers tended to describe a 4-ounce cup filled to the 2-ounce line as half full if it was previously empty but described it as half empty if it was previously full. Similar results were found when speakers could describe the outcome of a medical treatment in terms of either mortality or survival (e.g., 25% die vs. 75% survive). Two additional experiments showed that listeners made accurate inferences about speakers' reference points on the basis of the selected frame (e.g., if a speaker described a cup as half empty, listeners inferred that the cup used to be full). Taken together, the data suggest that frames reliably convey implicit information in addition to their explicit content, which helps explain why framing effects are so robust.

  10. Feasibility and effectiveness of an automated bilingual text message intervention for weight loss: pilot study.

    PubMed

    Kolodziejczyk, Julia K; Norman, Gregory J; Barrera-Ng, Angelica; Dillon, Lindsay; Marshall, Simon; Arredondo, Elva; Rock, Cheryl L; Raab, Fred; Griswold, William G; Sullivan, Mark; Patrick, Kevin

    2013-11-06

    Little is known about the feasibility and acceptability of tailored text message based weight loss programs for English and Spanish-language speakers. This pilot study evaluated the feasibility, acceptability, and estimated impact of a tailored text message based weight loss program for English and Spanish-language speakers. The purpose of this pilot study was to inform the development of a full-scale randomized trial. There were 20 overweight or obese participants (mean age 40.10, SD 8.05; 8/20, 40% male; 9/20, 45% Spanish-speakers) that were recruited in San Diego, California, from March to May 2011 and evaluated in a one-group pre/post clinical trial. For 8 weeks, participants received and responded to 3-5 text messages daily sent from a fully automated text messaging system. They also received printed weight loss materials and brief 10-15 minute weekly counseling calls. To estimate the impact of the program, the primary outcome was weight (kg) measured during face-to-face measurement visits by trained research staff. Pre and post differences in weight were analyzed with a one-way repeated measures analysis of variance. Differences by language preference at both time points were analyzed with t tests. Body mass index and weight management behaviors also were examined. Feasibility and acceptability were determined by recruitment success, adherence (ie, percentage of replies to interactive text messages and attrition), and participant satisfaction. Participants who completed the final assessment (N=18) decreased body weight by 1.85 kg (F1,17=10.80, P=.004, CI∆ 0.66-3.03, η(2)=0.39). At both time points, there were no differences in weight by language preference. Participants responded to 88.04% (986/1120) of interactive text messages, attrition rate was 10% (2/20), and 94% (19/20) of participants reported satisfaction with the program. This fully automated text message based weight program was feasible with English and Spanish-speakers and may have promoted modest weight loss over an 8-week period. Clinicaltrials.gov NCT01171586; http://clinicaltrials.gov/ct2/show/NCT01171586 (Archived by WebCite at http://www.webcitation.org/6Ksr6dl7n).

  11. Attitudes of Jordanian University Students towards Using Online Chat Discourse with Native Speakers of English for Improving Their Language Proficiency

    ERIC Educational Resources Information Center

    Mahfouz, Safi M.; Ihmeideh, Fathi M.

    2009-01-01

    This study aims to investigate Jordanian university students' attitudes towards using video and text chat discourse with anonymous native speakers of English to improve their English proficiency. To achieve this aim, a questionnaire was designed. The study sample consisted of 320 university students enrolled in two Jordanian universities. Results…

  12. Native Speakers of Arabic and ESL Texts: Evidence for the Transfer of Written Word Identification Processes

    ERIC Educational Resources Information Center

    Hayes-Harb, Rachel

    2006-01-01

    English as a second language (ESL) teachers have long noted that native speakers of Arabic exhibit exceptional difficulty with English reading comprehension (e.g., Thompson-Panos & Thomas-Ruzic, 1983). Most existing work in this area has looked to higher level aspects of reading such as familiarity with discourse structure and cultural knowledge…

  13. Non-Native Speaker Interaction Management Strategies in a Network-Based Virtual Environment

    ERIC Educational Resources Information Center

    Peterson, Mark

    2008-01-01

    This article investigates the dyad-based communication of two groups of non-native speakers (NNSs) of English involved in real time interaction in a type of text-based computer-mediated communication (CMC) tool known as a MOO. The object of this semester long study was to examine the ways in which the subjects managed their L2 interaction during…

  14. Native-Speaker/Non-Native-Speaker Discourse in the MOO: Topic Negotiation and Initiation in a Synchronous Text-Based Environment

    ERIC Educational Resources Information Center

    Schwienhorst, Klaus

    2004-01-01

    A number of researchers in computer-mediated communication have pointed towards its potential to stimulate learner participation and engagement in the classroom. However, in many cases only anecdotal reports were provided. In addition, it is unclear whether the pedagogical set-up or the technology involved is responsible for changes in learner…

  15. Optimization of multilayer neural network parameters for speaker recognition

    NASA Astrophysics Data System (ADS)

    Tovarek, Jaromir; Partila, Pavol; Rozhon, Jan; Voznak, Miroslav; Skapa, Jan; Uhrin, Dominik; Chmelikova, Zdenka

    2016-05-01

    This article discusses the impact of multilayer neural network parameters for speaker identification. The main task of speaker identification is to find a specific person in the known set of speakers. It means that the voice of an unknown speaker (wanted person) belongs to a group of reference speakers from the voice database. One of the requests was to develop the text-independent system, which means to classify wanted person regardless of content and language. Multilayer neural network has been used for speaker identification in this research. Artificial neural network (ANN) needs to set parameters like activation function of neurons, steepness of activation functions, learning rate, the maximum number of iterations and a number of neurons in the hidden and output layers. ANN accuracy and validation time are directly influenced by the parameter settings. Different roles require different settings. Identification accuracy and ANN validation time were evaluated with the same input data but different parameter settings. The goal was to find parameters for the neural network with the highest precision and shortest validation time. Input data of neural networks are a Mel-frequency cepstral coefficients (MFCC). These parameters describe the properties of the vocal tract. Audio samples were recorded for all speakers in a laboratory environment. Training, testing and validation data set were split into 70, 15 and 15 %. The result of the research described in this article is different parameter setting for the multilayer neural network for four speakers.

  16. Inclusion, Affection, Control: The Pragmatics of Intergenerational Communication.

    ERIC Educational Resources Information Center

    Hess, Lucille J.; Hess, Richard C.

    Personal intent and discourse considerations play an important role in understanding the nature of a conversation between a youth and an elderly person. Each participant makes assumptions about the listener's knowledge and ability to communicate effectively. The way a speaker reacts to the other participant depends upon the speaker's own…

  17. Processing subject-verb agreement in a second language depends on proficiency

    PubMed Central

    Hoshino, Noriko; Dussias, Paola E.; Kroll, Judith F.

    2010-01-01

    Subject-verb agreement is a computation that is often difficult to execute perfectly in the first language (L1) and even more difficult to produce skillfully in a second language (L2). In this study, we examined the way in which bilingual speakers complete sentence fragments in a manner that reflects access to both grammatical and conceptual number. In two experiments, we show that bilingual speakers are sensitive to both grammatical and conceptual number in the L1 and grammatical number agreement in the L2. However, only highly proficient bilinguals are also sensitive to conceptual number in the L2. The results suggest that the extent to which speakers are able to exploit conceptual information during speech planning depends on the level of language proficiency. PMID:20640178

  18. Syntactic learning by mere exposure - An ERP study in adult learners

    PubMed Central

    Mueller, Jutta L; Oberecker, Regine; Friederici, Angela D

    2009-01-01

    Background Artificial language studies have revealed the remarkable ability of humans to extract syntactic structures from a continuous sound stream by mere exposure. However, it remains unclear whether the processes acquired in such tasks are comparable to those applied during normal language processing. The present study compares the ERPs to auditory processing of simple Italian sentences in native and non-native speakers after brief exposure to Italian sentences of a similar structure. The sentences contained a non-adjacent dependency between an auxiliary and the morphologically marked suffix of the verb. Participants were presented four alternating learning and testing phases. During learning phases only correct sentences were presented while during testing phases 50 percent of the sentences contained a grammatical violation. Results The non-native speakers successfully learned the dependency and displayed an N400-like negativity and a subsequent anteriorily distributed positivity in response to rule violations. The native Italian group showed an N400 followed by a P600 effect. Conclusion The presence of the P600 suggests that native speakers applied a grammatical rule. In contrast, non-native speakers appeared to use a lexical form-based processing strategy. Thus, the processing mechanisms acquired in the language learning task were only partly comparable to those applied by competent native speakers. PMID:19640301

  19. Syntactic learning by mere exposure--an ERP study in adult learners.

    PubMed

    Mueller, Jutta L; Oberecker, Regine; Friederici, Angela D

    2009-07-29

    Artificial language studies have revealed the remarkable ability of humans to extract syntactic structures from a continuous sound stream by mere exposure. However, it remains unclear whether the processes acquired in such tasks are comparable to those applied during normal language processing. The present study compares the ERPs to auditory processing of simple Italian sentences in native and non-native speakers after brief exposure to Italian sentences of a similar structure. The sentences contained a non-adjacent dependency between an auxiliary and the morphologically marked suffix of the verb. Participants were presented four alternating learning and testing phases. During learning phases only correct sentences were presented while during testing phases 50 percent of the sentences contained a grammatical violation. The non-native speakers successfully learned the dependency and displayed an N400-like negativity and a subsequent anteriorily distributed positivity in response to rule violations. The native Italian group showed an N400 followed by a P600 effect. The presence of the P600 suggests that native speakers applied a grammatical rule. In contrast, non-native speakers appeared to use a lexical form-based processing strategy. Thus, the processing mechanisms acquired in the language learning task were only partly comparable to those applied by competent native speakers.

  20. Arctic Visiting Speakers Series (AVS)

    NASA Astrophysics Data System (ADS)

    Fox, S. E.; Griswold, J.

    2011-12-01

    The Arctic Visiting Speakers (AVS) Series funds researchers and other arctic experts to travel and share their knowledge in communities where they might not otherwise connect. Speakers cover a wide range of arctic research topics and can address a variety of audiences including K-12 students, graduate and undergraduate students, and the general public. Host applications are accepted on an on-going basis, depending on funding availability. Applications need to be submitted at least 1 month prior to the expected tour dates. Interested hosts can choose speakers from an online Speakers Bureau or invite a speaker of their choice. Preference is given to individuals and organizations to host speakers that reach a broad audience and the general public. AVS tours are encouraged to span several days, allowing ample time for interactions with faculty, students, local media, and community members. Applications for both domestic and international visits will be considered. Applications for international visits should involve participation of more than one host organization and must include either a US-based speaker or a US-based organization. This is a small but important program that educates the public about Arctic issues. There have been 27 tours since 2007 that have impacted communities across the globe including: Gatineau, Quebec Canada; St. Petersburg, Russia; Piscataway, New Jersey; Cordova, Alaska; Nuuk, Greenland; Elizabethtown, Pennsylvania; Oslo, Norway; Inari, Finland; Borgarnes, Iceland; San Francisco, California and Wolcott, Vermont to name a few. Tours have included lectures to K-12 schools, college and university students, tribal organizations, Boy Scout troops, science center and museum patrons, and the general public. There are approximately 300 attendees enjoying each AVS tour, roughly 4100 people have been reached since 2007. The expectations for each tour are extremely manageable. Hosts must submit a schedule of events and a tour summary to be posted online. Hosts must acknowledge the National Science Foundation Office of Polar Programs and ARCUS in all promotional materials. Host agrees to send ARCUS photographs, fliers, and if possible a video of the main lecture. Host and speaker agree to collect data on the number of attendees in each audience to submit as part of a post-tour evaluation. The grants can generally cover all the expenses of a tour, depending on the location. A maximum of 2,000 will be provided for the travel related expenses of a speaker on a domestic visit. A maxiμm of 2,500 will be provided for the travel related expenses of a speaker on an international visit. Each speaker will receive an honorarium of $300.

  1. Can you hear my age? Influences of speech rate and speech spontaneity on estimation of speaker age

    PubMed Central

    Skoog Waller, Sara; Eriksson, Mårten; Sörqvist, Patrik

    2015-01-01

    Cognitive hearing science is mainly about the study of how cognitive factors contribute to speech comprehension, but cognitive factors also partake in speech processing to infer non-linguistic information from speech signals, such as the intentions of the talker and the speaker’s age. Here, we report two experiments on age estimation by “naïve” listeners. The aim was to study how speech rate influences estimation of speaker age by comparing the speakers’ natural speech rate with increased or decreased speech rate. In Experiment 1, listeners were presented with audio samples of read speech from three different speaker age groups (young, middle aged, and old adults). They estimated the speakers as younger when speech rate was faster than normal and as older when speech rate was slower than normal. This speech rate effect was slightly greater in magnitude for older (60–65 years) speakers in comparison with younger (20–25 years) speakers, suggesting that speech rate may gain greater importance as a perceptual age cue with increased speaker age. This pattern was more pronounced in Experiment 2, in which listeners estimated age from spontaneous speech. Faster speech rate was associated with lower age estimates, but only for older and middle aged (40–45 years) speakers. Taken together, speakers of all age groups were estimated as older when speech rate decreased, except for the youngest speakers in Experiment 2. The absence of a linear speech rate effect in estimates of younger speakers, for spontaneous speech, implies that listeners use different age estimation strategies or cues (possibly vocabulary) depending on the age of the speaker and the spontaneity of the speech. Potential implications for forensic investigations and other applied domains are discussed. PMID:26236259

  2. "I'm Very Not About the Law Part": Nonnative Speakers of English and the Miranda Warnings

    ERIC Educational Resources Information Center

    Pavlenko, Aneta

    2008-01-01

    This article presents a case study of a police interrogation of a nonnative speaker (NNS) of English. I show that the high linguistic and conceptual complexity of police cautions, such as the Miranda warnings, complicates understanding of these texts even by NNSs of English with a high level of interactional competence. I argue that the U.S.…

  3. International Symposium on Internationalization and Foreign Students Problems. Proceedings (Kobe, Japan, October 29-30, 1990).

    ERIC Educational Resources Information Center

    Kobe Univ., (Japan).

    The meeting reported here was held to address the issues related to the increasing numbers of foreign students in Japan (31,000 in 1989 and increasing annually by 20%). First, profiles of the speakers and the texts of their speeches are presented, as well as the comments made during the panel discussions and symposium sessions. Speakers from 11…

  4. A Functional Imaging Study of Self-Regulatory Capacities in Persons Who Stutter

    PubMed Central

    Liu, Jie; Wang, Zhishun; Huo, Yuankai; Davidson, Stephanie M.; Klahr, Kristin; Herder, Carl L.; Sikora, Chamonix O.; Peterson, Bradley S.

    2014-01-01

    Developmental stuttering is a disorder of speech fluency with an unknown pathogenesis. The similarity of its phenotype and natural history with other childhood neuropsychiatric disorders of frontostriatal pathology suggests that stuttering may have a closely related pathogenesis. We investigated in this study the potential involvement of frontostriatal circuits in developmental stuttering. We collected functional magnetic resonance imaging data from 46 persons with stuttering and 52 fluent controls during performance of the Simon Spatial Incompatibility Task. We examined differences between the two groups of blood-oxygen-level-dependent activation associated with two neural processes, the resolution of cognitive conflict and the context-dependent adaptation to changes in conflict. Stuttering speakers and controls did not differ on behavioral performance on the task. In the presence of conflict-laden stimuli, however, stuttering speakers activated more strongly the cingulate cortex, left anterior prefrontal cortex, right medial frontal cortex, left supplementary motor area, right caudate nucleus, and left parietal cortex. The magnitude of activation in the anterior cingulate cortex correlated inversely in stuttering speakers with symptom severity. Stuttering speakers also showed blunted activation during context-dependent adaptation in the left dorsolateral prefrontal cortex, a brain region that mediates cross-temporal contingencies. Frontostriatal hyper-responsivity to conflict resembles prior findings in other disorders of frontostriatal pathology, and therefore likely represents a general mechanism supporting functional compensation for an underlying inefficiency of neural processing in these circuits. The reduced activation of dorsolateral prefrontal cortex likely represents the inadequate readiness of stuttering speakers to execute a sequence of motor responses. PMID:24587104

  5. A language-familiarity effect for speaker discrimination without comprehension.

    PubMed

    Fleming, David; Giordano, Bruno L; Caldara, Roberto; Belin, Pascal

    2014-09-23

    The influence of language familiarity upon speaker identification is well established, to such an extent that it has been argued that "Human voice recognition depends on language ability" [Perrachione TK, Del Tufo SN, Gabrieli JDE (2011) Science 333(6042):595]. However, 7-mo-old infants discriminate speakers of their mother tongue better than they do foreign speakers [Johnson EK, Westrek E, Nazzi T, Cutler A (2011) Dev Sci 14(5):1002-1011] despite their limited speech comprehension abilities, suggesting that speaker discrimination may rely on familiarity with the sound structure of one's native language rather than the ability to comprehend speech. To test this hypothesis, we asked Chinese and English adult participants to rate speaker dissimilarity in pairs of sentences in English or Mandarin that were first time-reversed to render them unintelligible. Even in these conditions a language-familiarity effect was observed: Both Chinese and English listeners rated pairs of native-language speakers as more dissimilar than foreign-language speakers, despite their inability to understand the material. Our data indicate that the language familiarity effect is not based on comprehension but rather on familiarity with the phonology of one's native language. This effect may stem from a mechanism analogous to the "other-race" effect in face recognition.

  6. Cost-sensitive learning for emotion robust speaker recognition.

    PubMed

    Li, Dongdong; Yang, Yingchun; Dai, Weihui

    2014-01-01

    In the field of information security, voice is one of the most important parts in biometrics. Especially, with the development of voice communication through the Internet or telephone system, huge voice data resources are accessed. In speaker recognition, voiceprint can be applied as the unique password for the user to prove his/her identity. However, speech with various emotions can cause an unacceptably high error rate and aggravate the performance of speaker recognition system. This paper deals with this problem by introducing a cost-sensitive learning technology to reweight the probability of test affective utterances in the pitch envelop level, which can enhance the robustness in emotion-dependent speaker recognition effectively. Based on that technology, a new architecture of recognition system as well as its components is proposed in this paper. The experiment conducted on the Mandarin Affective Speech Corpus shows that an improvement of 8% identification rate over the traditional speaker recognition is achieved.

  7. Cost-Sensitive Learning for Emotion Robust Speaker Recognition

    PubMed Central

    Li, Dongdong; Yang, Yingchun

    2014-01-01

    In the field of information security, voice is one of the most important parts in biometrics. Especially, with the development of voice communication through the Internet or telephone system, huge voice data resources are accessed. In speaker recognition, voiceprint can be applied as the unique password for the user to prove his/her identity. However, speech with various emotions can cause an unacceptably high error rate and aggravate the performance of speaker recognition system. This paper deals with this problem by introducing a cost-sensitive learning technology to reweight the probability of test affective utterances in the pitch envelop level, which can enhance the robustness in emotion-dependent speaker recognition effectively. Based on that technology, a new architecture of recognition system as well as its components is proposed in this paper. The experiment conducted on the Mandarin Affective Speech Corpus shows that an improvement of 8% identification rate over the traditional speaker recognition is achieved. PMID:24999492

  8. ``The perceptual bases of speaker identity'' revisited

    NASA Astrophysics Data System (ADS)

    Voiers, William D.

    2003-10-01

    A series of experiments begun 40 years ago [W. D. Voiers, J. Acoust. Soc. Am. 36, 1065-1073 (1964)] was concerned with identifying the perceived voice traits (PVTs) on which human recognition of voices depends. It culminated with the development of a voice taxonomy based on 20 PVTs and a set of highly reliable rating scales for classifying voices with respect to those PVTs. The development of a perceptual voice taxonomy was motivated by the need for a practical method of evaluating speaker recognizability in voice communication systems. The Diagnostic Speaker Recognition Test (DSRT) evaluates the effects of systems on speaker recognizability as reflected in changes in the inter-listener reliability of voice ratings on the 20 PVTs. The DSRT thus provides a qualitative, as well as quantitative, evaluation of the effects of a system on speaker recognizability. A fringe benefit of this project is PVT rating data for a sample of 680 voices. [Work partially supported by USAFRL.

  9. The prevalence of synaesthesia depends on early language learning.

    PubMed

    Watson, Marcus R; Chromý, Jan; Crawford, Lyle; Eagleman, David M; Enns, James T; Akins, Kathleen A

    2017-02-01

    According to one theory, synaesthesia develops, or is preserved, because it helps children learn. If so, it should be more common among adults who faced greater childhood learning challenges. In the largest survey of synaesthesia to date, the incidence of synaesthesia was compared among native speakers of languages with transparent (easier) and opaque (more difficult) orthographies. Contrary to our prediction, native speakers of Czech (transparent) were more likely to be synaesthetes than native speakers of English (opaque). However, exploratory analyses suggested that this was because more Czechs learned non-native second languages, which was strongly associated with synaesthesia, consistent with the learning hypothesis. Furthermore, the incidence of synaesthesia among speakers of opaque languages was double that among speakers of transparent languages other than Czech, also consistent with the learning hypothesis. These findings contribute to an emerging understanding of synaesthetic development as a complex and lengthy process with multiple causal influences. Copyright © 2016. Published by Elsevier Inc.

  10. English as a Second Language for Adults. Discussion Paper 04/79.

    ERIC Educational Resources Information Center

    Selman, Mary

    Because of a growing community of non-English speakers in British Columbia, there is an urgent need for effective teaching programs in English as a Second Language (ESL). Non-English speakers frequently face educational deprivation, difficulty in using their skills and in finding employment, dependency on government assistance, and, if children,…

  11. The object of my desire: Five-year-olds rapidly reason about a speaker's desire during referential communication.

    PubMed

    San Juan, Valerie; Chambers, Craig G; Berman, Jared; Humphry, Chelsea; Graham, Susan A

    2017-10-01

    Two experiments examined whether 5-year-olds draw inferences about desire outcomes that constrain their online interpretation of an utterance. Children were informed of a speaker's positive (Experiment 1) or negative (Experiment 2) desire to receive a specific toy as a gift before hearing a referentially ambiguous statement ("That's my present") spoken with either a happy or sad voice. After hearing the speaker express a positive desire, children (N=24) showed an implicit (i.e., eye gaze) and explicit ability to predict reference to the desired object when the speaker sounded happy, but they showed only implicit consideration of the alternate object when the speaker sounded sad. After hearing the speaker express a negative desire, children (N=24) used only happy prosodic cues to predict the intended referent of the statement. Taken together, the findings indicate that the efficiency with which 5-year-olds integrate desire reasoning with language processing depends on the emotional valence of the speaker's voice but not on the type of desire representations (i.e., positive vs. negative) that children must reason about online. Copyright © 2017 Elsevier Inc. All rights reserved.

  12. ASTP Onboard Voice Transcription

    NASA Technical Reports Server (NTRS)

    1975-01-01

    The transcription is presented of the Apollo-Soyuz Test Project voice communications as recorded on the command module data storage equipment. Data from this recorder are telemetered (dumped) to Space Tracking and Data Network sites for retransmission to the Johnson Space Center. The transcript is divided into three columns -- time, speaker, and text. The Greenwich mean time column consists of three two-digit numbers representing hours, minutes, and seconds (e.g., 22 34 14) for the Julian dates shown at the top of the page on which a new day begins. The speaker column indicates the source of a transmission; the text column contains the verbatim transcript of the communications.

  13. The Influence of Orthography on the Production of Alphabetic, Second-Language Allophones by Speakers of a Non-alphabetic Language.

    PubMed

    Han, Jeong-Im; Kim, Joo-Yeon

    2017-08-01

    This study investigated the influence of orthographic information on the production of allophones in a second language (L2). Two proficiency levels of native Mandarin speakers learned novel Korean words with potential variants of /h/ based on auditory stimuli, and then they were provided various types of spellings for the variants, including the letters for [[Formula: see text

  14. Euclidean Distances as measures of speaker similarity including identical twin pairs: A forensic investigation using source and filter voice characteristics.

    PubMed

    San Segundo, Eugenia; Tsanas, Athanasios; Gómez-Vilda, Pedro

    2017-01-01

    There is a growing consensus that hybrid approaches are necessary for successful speaker characterization in Forensic Speaker Comparison (FSC); hence this study explores the forensic potential of voice features combining source and filter characteristics. The former relate to the action of the vocal folds while the latter reflect the geometry of the speaker's vocal tract. This set of features have been extracted from pause fillers, which are long enough for robust feature estimation while spontaneous enough to be extracted from voice samples in real forensic casework. Speaker similarity was measured using standardized Euclidean Distances (ED) between pairs of speakers: 54 different-speaker (DS) comparisons, 54 same-speaker (SS) comparisons and 12 comparisons between monozygotic twins (MZ). Results revealed that the differences between DS and SS comparisons were significant in both high quality and telephone-filtered recordings, with no false rejections and limited false acceptances; this finding suggests that this set of voice features is highly speaker-dependent and therefore forensically useful. Mean ED for MZ pairs lies between the average ED for SS comparisons and DS comparisons, as expected according to the literature on twin voices. Specific cases of MZ speakers with very high ED (i.e. strong dissimilarity) are discussed in the context of sociophonetic and twin studies. A preliminary simplification of the Vocal Profile Analysis (VPA) Scheme is proposed, which enables the quantification of voice quality features in the perceptual assessment of speaker similarity, and allows for the calculation of perceptual-acoustic correlations. The adequacy of z-score normalization for this study is also discussed, as well as the relevance of heat maps for detecting the so-called phantoms in recent approaches to the biometric menagerie. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.

  15. Proficiency and Working Memory Based Explanations for Nonnative Speakers' Sensitivity to Agreement in Sentence Processing

    ERIC Educational Resources Information Center

    Coughlin, Caitlin E.; Tremblay, Annie

    2013-01-01

    This study examines the roles of proficiency and working memory (WM) capacity in second-/foreign-language (L2) learners' processing of agreement morphology. It investigates the processing of grammatical and ungrammatical short- and long-distance number agreement dependencies by native English speakers at two proficiencies in French, and the…

  16. Downriver Indians' Legends.

    ERIC Educational Resources Information Center

    Bennett, Ruth, Ed.; Exline, Jesse

    Yurok Indian legends in Yurok Unifon text include English translations of the entire texts in order to produce fluent reading for English speakers and a continuous text for Yurok readers. Although corresponding sentences are numbered, translation is not word-for-word or sentence-for-sentence. The five stories refer to a time when animals could…

  17. A Cross-Language Study of Acoustic Predictors of Speech Intelligibility in Individuals With Parkinson's Disease

    PubMed Central

    Choi, Yaelin

    2017-01-01

    Purpose The present study aimed to compare acoustic models of speech intelligibility in individuals with the same disease (Parkinson's disease [PD]) and presumably similar underlying neuropathologies but with different native languages (American English [AE] and Korean). Method A total of 48 speakers from the 4 speaker groups (AE speakers with PD, Korean speakers with PD, healthy English speakers, and healthy Korean speakers) were asked to read a paragraph in their native languages. Four acoustic variables were analyzed: acoustic vowel space, voice onset time contrast scores, normalized pairwise variability index, and articulation rate. Speech intelligibility scores were obtained from scaled estimates of sentences extracted from the paragraph. Results The findings indicated that the multiple regression models of speech intelligibility were different in Korean and AE, even with the same set of predictor variables and with speakers matched on speech intelligibility across languages. Analysis of the descriptive data for the acoustic variables showed the expected compression of the vowel space in speakers with PD in both languages, lower normalized pairwise variability index scores in Korean compared with AE, and no differences within or across language in articulation rate. Conclusions The results indicate that the basis of an intelligibility deficit in dysarthria is likely to depend on the native language of the speaker and listener. Additional research is required to explore other potential predictor variables, as well as additional language comparisons to pursue cross-linguistic considerations in classification and diagnosis of dysarthria types. PMID:28821018

  18. I "hear" what you're "saying": Auditory perceptual simulation, reading speed, and reading comprehension.

    PubMed

    Zhou, Peiyun; Christianson, Kiel

    2016-01-01

    Auditory perceptual simulation (APS) during silent reading refers to situations in which the reader actively simulates the voice of a character or other person depicted in a text. In three eye-tracking experiments, APS effects were investigated as people read utterances attributed to a native English speaker, a non-native English speaker, or no speaker at all. APS effects were measured via online eye movements and offline comprehension probes. Results demonstrated that inducing APS during silent reading resulted in observable differences in reading speed when readers simulated the speech of faster compared to slower speakers and compared to silent reading without APS. Social attitude survey results indicated that readers' attitudes towards the native and non-native speech did not consistently influence APS-related effects. APS of both native speech and non-native speech increased reading speed, facilitated deeper, less good-enough sentence processing, and improved comprehension compared to normal silent reading.

  19. The Effects of Syntactically Parsed Text Formats on Intensive Reading in EFL

    ERIC Educational Resources Information Center

    Herbert, John C.

    2014-01-01

    Separating text into meaningful language chunks, as with visual-syntactic text formatting, helps readers to process text more easily and language learners to recognize grammar and syntax patterns more quickly. Evidence of this exists in studies on native and non-native English speakers. However, recent studies question the roll of VSTF in certain…

  20. Metalinguistic awareness and reading performance: a cross language comparison.

    PubMed

    Ibrahim, Raphiq; Eviatar, Zohar; Aharon-Peretz, Judith

    2007-07-01

    The study examined two questions: (1) do the greater phonological awareness skills of billinguals affect reading performance; (2) to what extent do the orthographic characteristics of a language influence reading performance and how does this interact with the effects of phonological awareness. We estimated phonological metalinguistic abilities and reading measures in three groups of first graders: monolingual Hebrew speakers, bilingual Russian-Hebrew speakers, and Arabic-speaking children. We found that language experience affects phonological awareness, as both Russian-Hebrew bilinguals and the Arabic speakers achieved higher scores on metalinguistic tests than Hebrew speakers. Orthography affected reading measures and their correlation with phonological abilitites. Children reading Hebrew showed better text reading ability and significant correlations between phonological awareness and reading scores. Children reading Arabic showed a slight advantage in single word and nonword reading over the two Hebrew reading groups, and very weak relationships between phonological abilities and reading performance. We conclude that native Arabic speakers have more difficulty in processing Arabic orthography than Hebrew monolinguals and bilinguals have in processing Hebrew orthography, and suggest that this is due to the additional visual complexity of Arabic orthography.

  1. Visual abilities are important for auditory-only speech recognition: evidence from autism spectrum disorder.

    PubMed

    Schelinski, Stefanie; Riedel, Philipp; von Kriegstein, Katharina

    2014-12-01

    In auditory-only conditions, for example when we listen to someone on the phone, it is essential to fast and accurately recognize what is said (speech recognition). Previous studies have shown that speech recognition performance in auditory-only conditions is better if the speaker is known not only by voice, but also by face. Here, we tested the hypothesis that such an improvement in auditory-only speech recognition depends on the ability to lip-read. To test this we recruited a group of adults with autism spectrum disorder (ASD), a condition associated with difficulties in lip-reading, and typically developed controls. All participants were trained to identify six speakers by name and voice. Three speakers were learned by a video showing their face and three others were learned in a matched control condition without face. After training, participants performed an auditory-only speech recognition test that consisted of sentences spoken by the trained speakers. As a control condition, the test also included speaker identity recognition on the same auditory material. The results showed that, in the control group, performance in speech recognition was improved for speakers known by face in comparison to speakers learned in the matched control condition without face. The ASD group lacked such a performance benefit. For the ASD group auditory-only speech recognition was even worse for speakers known by face compared to speakers not known by face. In speaker identity recognition, the ASD group performed worse than the control group independent of whether the speakers were learned with or without face. Two additional visual experiments showed that the ASD group performed worse in lip-reading whereas face identity recognition was within the normal range. The findings support the view that auditory-only communication involves specific visual mechanisms. Further, they indicate that in ASD, speaker-specific dynamic visual information is not available to optimize auditory-only speech recognition. Copyright © 2014 Elsevier Ltd. All rights reserved.

  2. An Event-Related Potential (ERP) Investigation of Filler-Gap Processing in Native and Second Language Speakers

    ERIC Educational Resources Information Center

    Dallas, Andrea; DeDe, Gayle; Nicol, Janet

    2013-01-01

    The current study employed a neuro-imaging technique, Event-Related Potentials (ERP), to investigate real-time processing of sentences containing filler-gap dependencies by late-learning speakers of English as a second language (L2) with a Chinese native language background. An individual differences approach was also taken to examine the role of…

  3. Use of Speaker Intent and Grammatical Cues in Fast-Mapping by Adolescents with Down Syndrome

    ERIC Educational Resources Information Center

    McDuffie, Andrea S.; Sindberg, Heidi A.; Hesketh, Linda J.; Chapman, Robin S.

    2007-01-01

    Purpose: The authors asked whether adolescents with Down syndrome (DS) could fast-map novel nouns and verbs when word learning depended on using the speaker's pragmatic or syntactic cues. Compared with typically developing (TD) comparison children, the authors predicted that syntactic cues would prove harder for the group with DS to use and that…

  4. Utterance selection model of language change

    NASA Astrophysics Data System (ADS)

    Baxter, G. J.; Blythe, R. A.; Croft, W.; McKane, A. J.

    2006-04-01

    We present a mathematical formulation of a theory of language change. The theory is evolutionary in nature and has close analogies with theories of population genetics. The mathematical structure we construct similarly has correspondences with the Fisher-Wright model of population genetics, but there are significant differences. The continuous time formulation of the model is expressed in terms of a Fokker-Planck equation. This equation is exactly soluble in the case of a single speaker and can be investigated analytically in the case of multiple speakers who communicate equally with all other speakers and give their utterances equal weight. Whilst the stationary properties of this system have much in common with the single-speaker case, time-dependent properties are richer. In the particular case where linguistic forms can become extinct, we find that the presence of many speakers causes a two-stage relaxation, the first being a common marginal distribution that persists for a long time as a consequence of ultimate extinction being due to rare fluctuations.

  5. Mamanwa Texts. Studies in Philippine Linguistics. Supplementary Series: Philippine Texts, No. 5.

    ERIC Educational Resources Information Center

    Miller, Helen, Comp.; Miller, Jeanne, Comp.

    A group of 28 texts in Mamanwa, an Austronesian language of the Visayan family of the Southern Philippines, is presented. They were told by native speakers and transcribed in Mamanwa and English. Three types of discourse are represented: narrative (both traditional and factual); procedural/explanatory; and hortatory. The main cultural themes…

  6. Speech enhancement based on neural networks improves speech intelligibility in noise for cochlear implant users.

    PubMed

    Goehring, Tobias; Bolner, Federico; Monaghan, Jessica J M; van Dijk, Bas; Zarowski, Andrzej; Bleeck, Stefan

    2017-02-01

    Speech understanding in noisy environments is still one of the major challenges for cochlear implant (CI) users in everyday life. We evaluated a speech enhancement algorithm based on neural networks (NNSE) for improving speech intelligibility in noise for CI users. The algorithm decomposes the noisy speech signal into time-frequency units, extracts a set of auditory-inspired features and feeds them to the neural network to produce an estimation of which frequency channels contain more perceptually important information (higher signal-to-noise ratio, SNR). This estimate is used to attenuate noise-dominated and retain speech-dominated CI channels for electrical stimulation, as in traditional n-of-m CI coding strategies. The proposed algorithm was evaluated by measuring the speech-in-noise performance of 14 CI users using three types of background noise. Two NNSE algorithms were compared: a speaker-dependent algorithm, that was trained on the target speaker used for testing, and a speaker-independent algorithm, that was trained on different speakers. Significant improvements in the intelligibility of speech in stationary and fluctuating noises were found relative to the unprocessed condition for the speaker-dependent algorithm in all noise types and for the speaker-independent algorithm in 2 out of 3 noise types. The NNSE algorithms used noise-specific neural networks that generalized to novel segments of the same noise type and worked over a range of SNRs. The proposed algorithm has the potential to improve the intelligibility of speech in noise for CI users while meeting the requirements of low computational complexity and processing delay for application in CI devices. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.

  7. Facilitating Comprehension of Non-Native English Speakers during Lectures in English with STR-Texts

    ERIC Educational Resources Information Center

    Shadiev, Rustam; Wu, Ting-Ting; Huang, Yueh-Min

    2018-01-01

    We provided texts generated by speech-to text-recognition (STR) technology for non-native English speaking students during lectures in English in order to test whether STR-texts were useful for enhancing students' comprehension of lectures. To this end, we carried out an experiment in which 60 participants were randomly assigned to a control group…

  8. Revisiting the role of language in spatial cognition: Categorical perception of spatial relations in English and Korean speakers.

    PubMed

    Holmes, Kevin J; Moty, Kelsey; Regier, Terry

    2017-12-01

    The spatial relation of support has been regarded as universally privileged in nonlinguistic cognition and immune to the influence of language. English, but not Korean, obligatorily distinguishes support from nonsupport via basic spatial terms. Despite this linguistic difference, previous research suggests that English and Korean speakers show comparable nonlinguistic sensitivity to the support/nonsupport distinction. Here, using a paradigm previously found to elicit cross-language differences in color discrimination, we provide evidence for a difference in sensitivity to support/nonsupport between native English speakers and native Korean speakers who were late English learners and tested in a context that privileged Korean. Whereas the former group showed categorical perception (CP) when discriminating spatial scenes capturing the support/nonsupport distinction, the latter did not. An additional group of native Korean speakers-relatively early English learners tested in an English-salient context-patterned with the native English speakers in showing CP for support/nonsupport. These findings suggest that obligatory marking of support/nonsupport in one's native language can affect nonlinguistic sensitivity to this distinction, contra earlier findings, but that such sensitivity may also depend on aspects of language background and the immediate linguistic context.

  9. Teaching the Order of Adjectives in the English Noun Phrase.

    ERIC Educational Resources Information Center

    Ney, James W.

    A number of studies on the order of adjectives in the English noun phrase are reviewed. Analysis of the studies and examples used in them indicates that almost any order of adjective seems to be possible depending on the intended meaning of the speaker or the situation in which the speaker frames an utterance. To see if in fact the ordering of…

  10. Discourse intonation and second language acquisition: Three genre-based studies

    NASA Astrophysics Data System (ADS)

    Wennerstrom, Ann Kristin

    1997-12-01

    This dissertation investigates intonation in the discourse of nonnative speakers of English. It is proposed that intonation functions as a grammar of cohesion, contributing to the coherence of the text. Based on a componential model of intonation adapted from Pierrehumbert and Hirshberg (1990), three empirical studies were conducted in different genres of spoken discourse: academic lectures, conversations, and oral narratives. Using computerized speech technology, excerpts of taped discourse were measured to determine how intonation associated with various constituents of text. All speakers were tested for overall English level on tests adapted from the SPEAK Test (ETS, 1985). Comparisons using native speaker data were also conducted. The first study investigated intonation in lectures given by Chinese teaching assistants. Multivariate analyses showed that intonation was a significant factor contributing to better scores on an exam of overall comprehensibility in English. The second study investigated the role of intonation in the turn-taking system in conversations between native and nonnative speakers of English. The final study considered emotional aspects of intonation in narratives, using the framework of Labov and Waletsky (1967). In sum, adult nonnative speakers can acquire intonation as part of their overall language development, although there is evidence against any specific order of acquisition. Intonation contributes to coherence by indicating the relationship between the current utterance and what is assumed to already be in participants' mental representations of the discourse. It also performs a segmentation function, denoting hierarchical relationships among utterances and/or turns. It is suggested that while pitch can be a resource in cross-cultural communication to show emotion and attitude, the grammatical aspects of intonation must be acquired gradually.

  11. A fundamental residue pitch perception bias for tone language speakers

    NASA Astrophysics Data System (ADS)

    Petitti, Elizabeth

    A complex tone composed of only higher-order harmonics typically elicits a pitch percept equivalent to the tone's missing fundamental frequency (f0). When judging the direction of residue pitch change between two such tones, however, listeners may have completely opposite perceptual experiences depending on whether they are biased to perceive changes based on the overall spectrum or the missing f0 (harmonic spacing). Individual differences in residue pitch change judgments are reliable and have been associated with musical experience and functional neuroanatomy. Tone languages put greater pitch processing demands on their speakers than non-tone languages, and we investigated whether these lifelong differences in linguistic pitch processing affect listeners' bias for residue pitch. We asked native tone language speakers and native English speakers to perform a pitch judgment task for two tones with missing fundamental frequencies. Given tone pairs with ambiguous pitch changes, listeners were asked to judge the direction of pitch change, where the direction of their response indicated whether they attended to the overall spectrum (exhibiting a spectral bias) or the missing f0 (exhibiting a fundamental bias). We found that tone language speakers are significantly more likely to perceive pitch changes based on the missing f0 than English speakers. These results suggest that tone-language speakers' privileged experience with linguistic pitch fundamentally tunes their basic auditory processing.

  12. Auditory perceptual simulation: Simulating speech rates or accents?

    PubMed

    Zhou, Peiyun; Christianson, Kiel

    2016-07-01

    When readers engage in Auditory Perceptual Simulation (APS) during silent reading, they mentally simulate characteristics of voices attributed to a particular speaker or a character depicted in the text. Previous research found that auditory perceptual simulation of a faster native English speaker during silent reading led to shorter reading times that auditory perceptual simulation of a slower non-native English speaker. Yet, it was uncertain whether this difference was triggered by the different speech rates of the speakers, or by the difficulty of simulating an unfamiliar accent. The current study investigates this question by comparing faster Indian-English speech and slower American-English speech in the auditory perceptual simulation paradigm. Analyses of reading times of individual words and the full sentence reveal that the auditory perceptual simulation effect again modulated reading rate, and auditory perceptual simulation of the faster Indian-English speech led to faster reading rates compared to auditory perceptual simulation of the slower American-English speech. The comparison between this experiment and the data from Zhou and Christianson (2016) demonstrate further that the "speakers'" speech rates, rather than the difficulty of simulating a non-native accent, is the primary mechanism underlying auditory perceptual simulation effects. Copyright © 2016 Elsevier B.V. All rights reserved.

  13. Adaptive Communication: Languages with More Non-Native Speakers Tend to Have Fewer Word Forms

    PubMed Central

    Bentz, Christian; Verkerk, Annemarie; Kiela, Douwe; Hill, Felix; Buttery, Paula

    2015-01-01

    Explaining the diversity of languages across the world is one of the central aims of typological, historical, and evolutionary linguistics. We consider the effect of language contact-the number of non-native speakers a language has-on the way languages change and evolve. By analysing hundreds of languages within and across language families, regions, and text types, we show that languages with greater levels of contact typically employ fewer word forms to encode the same information content (a property we refer to as lexical diversity). Based on three types of statistical analyses, we demonstrate that this variance can in part be explained by the impact of non-native speakers on information encoding strategies. Finally, we argue that languages are information encoding systems shaped by the varying needs of their speakers. Language evolution and change should be modeled as the co-evolution of multiple intertwined adaptive systems: On one hand, the structure of human societies and human learning capabilities, and on the other, the structure of language. PMID:26083380

  14. Suppression and Working Memory in Auditory Comprehension of L2 Narratives: Evidence from Cross-Modal Priming.

    PubMed

    Wu, Shiyu; Ma, Zheng

    2016-10-01

    Using a cross-modal priming task, the present study explores whether Chinese-English bilinguals process goal related information during auditory comprehension of English narratives like native speakers. Results indicate that English native speakers adopted both mechanisms of suppression and enhancement to modulate the activation of goals and keep track of the "causal path" in narrative events and that L1 speakers with higher working memory (WM) capacity are more skilled at attenuating interference. L2 speakers, however, experienced the phenomenon of "facilitation-without-inhibition." Their difficulty in suppressing irrelevant information was related to their performance in the test of working memory capacity. For the L2 group with greater working memory capacity, the effects of both enhancement and suppression were found. These findings are discussed in light of a landscape model of L2 text comprehension which highlights the need for WM to be incorporated into comprehensive models of L2 processing as well as theories of SLA.

  15. Electrophysiology of subject-verb agreement mediated by speakers' gender.

    PubMed

    Hanulíková, Adriana; Carreiras, Manuel

    2015-01-01

    An important property of speech is that it explicitly conveys features of a speaker's identity such as age or gender. This event-related potential (ERP) study examined the effects of social information provided by a speaker's gender, i.e., the conceptual representation of gender, on subject-verb agreement. Despite numerous studies on agreement, little is known about syntactic computations generated by speaker characteristics extracted from the acoustic signal. Slovak is well suited to investigate this issue because it is a morphologically rich language in which agreement involves features for number, case, and gender. Grammaticality of a sentence can be evaluated by checking a speaker's gender as conveyed by his/her voice. We examined how conceptual information about speaker gender, which is not syntactic but rather social and pragmatic in nature, is interpreted for the computation of agreement patterns. ERP responses to verbs disagreeing with the speaker's gender (e.g., a sentence including a masculine verbal inflection spoken by a female person 'the neighbors were upset because I (∗)stoleMASC plums') elicited a larger early posterior negativity compared to correct sentences. When the agreement was purely syntactic and did not depend on the speaker's gender, a disagreement between a formally marked subject and the verb inflection (e.g., the womanFEM (∗)stoleMASC plums) resulted in a larger P600 preceded by a larger anterior negativity compared to the control sentences. This result is in line with proposals according to which the recruitment of non-syntactic information such as the gender of the speaker results in N400-like effects, while formally marked syntactic features lead to structural integration as reflected in a LAN/P600 complex.

  16. Assimilation and accommodation patterns in ventral occipitotemporal cortex in learning a second writing system

    PubMed Central

    Nelson, Jessica R.; Liu, Ying; Fiez, Julie; Perfetti, Charles A.

    2017-01-01

    Using fMRI, we compared the patterns of fusiform activity produced by viewing English and Chinese for readers who were either English speakers learning Chinese, or Chinese-English bilinguals. The pattern of fusiform activity depended on both the writing system and the reader’s native language. Native Chinese speakers fluent in English recruited bilateral fusiform areas when viewing both Chinese and English. English speakers learning Chinese, however, used heavily left-lateralized fusiform regions when viewing English, but recruited an additional right fusiform region for viewing Chinese. Thus, English learners of Chinese show an accommodation pattern, in which the reading network accommodates the new writing system by adding neural resources that support its specific graphic requirements. Chinese speakers show an assimilation pattern, in which the reading network established for L1 includes procedures sufficient for the graphic demands of L2 without major change. PMID:18381767

  17. Coupled Electro-Magneto-Mechanical-Acoustic Analysis Method Developed by Using 2D Finite Element Method for Flat Panel Speaker Driven by Magnetostrictive-Material-Based Actuator

    NASA Astrophysics Data System (ADS)

    Yoo, Byungjin; Hirata, Katsuhiro; Oonishi, Atsurou

    In this study, a coupled analysis method for flat panel speakers driven by giant magnetostrictive material (GMM) based actuator was developed. The sound field produced by a flat panel speaker that is driven by a GMM actuator depends on the vibration of the flat panel, this vibration is a result of magnetostriction property of the GMM. In this case, to predict the sound pressure level (SPL) in the audio-frequency range, it is necessary to take into account not only the magnetostriction property of the GMM but also the effect of eddy current and the vibration characteristics of the actuator and the flat panel. In this paper, a coupled electromagnetic-structural-acoustic analysis method is presented; this method was developed by using the finite element method (FEM). This analysis method is used to predict the performance of a flat panel speaker in the audio-frequency range. The validity of the analysis method is verified by comparing with the measurement results of a prototype speaker.

  18. Reaching Spanish-speaking smokers online: a 10-year worldwide research program

    PubMed Central

    Muñoz, Ricardo Felipe; Chen, Ken; Bunge, Eduardo Liniers; Bravin, Julia Isabela; Shaughnessy, Elizabeth Annelly; Pérez-Stable, Eliseo Joaquín

    2014-01-01

    Objective To describe a 10-year proof-of-concept smoking cessation research program evaluating the reach of online health interventions throughout the Americas. Methods Recruitment occurred from 2002–2011, primarily using Google.com AdWords. Over 6 million smokers from the Americas entered keywords related to smoking cessation; 57 882 smokers (15 912 English speakers and 41 970 Spanish speakers) were recruited into online self-help automated intervention studies. To examine disparities in utilization of methods to quit smoking, cessation aids used by English speakers and Spanish speakers were compared. To determine whether online interventions reduce disparities, abstinence rates were also compared. Finally, the reach of the intervention was illustrated for three large Spanish-speaking countries of the Americas—Argentina, Mexico, and Peru—and the United States of America. Results Few participants had utilized other methods to stop smoking before coming to the Internet site; most reported using no previous smoking cessation aids: 69.2% of Spanish speakers versus 51.8% of English speakers (P < 0.01). The most used method was nicotine gum, 13.9%. Nicotine dependence levels were similar to those reported for in-person smoking cessation trials. Overall observed quit rate for English speakers was 38.1% and for Spanish speakers, 37.0%; quit rates in which participants with missing data were considered to be smoking were 11.1% and 10.6%, respectively. Neither comparison was significantly different. Conclusions The systematic use of evidence-based Internet interventions for health problems could have a broad impact throughout the Americas, at little or no cost to individuals or to ministries of health. PMID:25211569

  19. Attentional influences on functional mapping of speech sounds in human auditory cortex.

    PubMed

    Obleser, Jonas; Elbert, Thomas; Eulitz, Carsten

    2004-07-21

    The speech signal contains both information about phonological features such as place of articulation and non-phonological features such as speaker identity. These are different aspects of the 'what'-processing stream (speaker vs. speech content), and here we show that they can be further segregated as they may occur in parallel but within different neural substrates. Subjects listened to two different vowels, each spoken by two different speakers. During one block, they were asked to identify a given vowel irrespectively of the speaker (phonological categorization), while during the other block the speaker had to be identified irrespectively of the vowel (speaker categorization). Auditory evoked fields were recorded using 148-channel magnetoencephalography (MEG), and magnetic source imaging was obtained for 17 subjects. During phonological categorization, a vowel-dependent difference of N100m source location perpendicular to the main tonotopic gradient replicated previous findings. In speaker categorization, the relative mapping of vowels remained unchanged but sources were shifted towards more posterior and more superior locations. These results imply that the N100m reflects the extraction of abstract invariants from the speech signal. This part of the processing is accomplished in auditory areas anterior to AI, which are part of the auditory 'what' system. This network seems to include spatially separable modules for identifying the phonological information and for associating it with a particular speaker that are activated in synchrony but within different regions, suggesting that the 'what' processing can be more adequately modeled by a stream of parallel stages. The relative activation of the parallel processing stages can be modulated by attentional or task demands.

  20. Text Messaging, Pragmatic Competence, and Affective Facilitation in the EFL Context: A Pilot Study

    ERIC Educational Resources Information Center

    Haggan, Madeline

    2010-01-01

    Since text messaging is a widely popular method of communication among young people, the paper tries to investigate whether it might have some practical application in the EFL classroom. Kuwaiti EFL students asked to render a mixture of text messages written by their peers and native English speakers into Standard English produced a large number…

  1. Semi-Spontaneous Oral Text Production: Measurements in Clinical Practice

    ERIC Educational Resources Information Center

    Lind, Marianne; Kristoffersen, Kristian Emil; Moen, Inger; Simonsen, Hanne Gram

    2009-01-01

    Functionally relevant assessment of the language production of speakers with aphasia should include assessment of connected speech production. Despite the ecological validity of everyday conversations, more controlled and monological types of texts may be easier to obtain and analyse in clinical practice. This article discusses some simple…

  2. A Formative Study: Inquiry and Informational Text with Fifth-Grade Bilinguals

    ERIC Educational Resources Information Center

    Moses, Lindsey

    2014-01-01

    This article includes the findings from a formative experiment implementing inquiry with informational texts in a fifth-grade bilingual classroom after the completion of state assessments. The pedagogical goals were focused on facilitating engaged reading and writing for native Spanish-speakers and building content knowledge and related academic…

  3. Introductory Punjabi.

    ERIC Educational Resources Information Center

    Bahri, Ujjal Singh; Walia, Paramjit Singh

    This introductory text in Punjabi (also spelled Panjabi) is intended primarily for those whose mother tongue is not Punjabi but are native speakers of other Indian languages. Some familiarity with the Punjabi cultural items is presupposed. The non-Indian may, however, also be able to use this text with profit since the lessons are graded. The…

  4. Comparing headphone and speaker effects on simulated driving.

    PubMed

    Nelson, T M; Nilsson, T H

    1990-12-01

    Twelve persons drove for three hours in an automobile simulator while listening to music at sound level 63dB over stereo headphones during one session and from a dashboard speaker during another session. They were required to steer a mountain highway, maintain a certain indicated speed, shift gears, and respond to occasional hazards. Steering and speed control were dependent on visual cues. The need to shift and the hazards were indicated by sound and vibration effects. With the headphones, the driver's average reaction time for the most complex task presented--shifting gears--was about one-third second longer than with the speaker. The use of headphones did not delay the development of subjective fatigue.

  5. Individual aptitude in Mandarin lexical tone perception predicts effectiveness of high-variability training

    PubMed Central

    Sadakata, Makiko; McQueen, James M.

    2014-01-01

    Although the high-variability training method can enhance learning of non-native speech categories, this can depend on individuals’ aptitude. The current study asked how general the effects of perceptual aptitude are by testing whether they occur with training materials spoken by native speakers and whether they depend on the nature of the to-be-learned material. Forty-five native Dutch listeners took part in a 5-day training procedure in which they identified bisyllabic Mandarin pseudowords (e.g., asa) pronounced with different lexical tone combinations. The training materials were presented to different groups of listeners at three levels of variability: low (many repetitions of a limited set of words recorded by a single speaker), medium (fewer repetitions of a more variable set of words recorded by three speakers), and high (similar to medium but with five speakers). Overall, variability did not influence learning performance, but this was due to an interaction with individuals’ perceptual aptitude: increasing variability hindered improvements in performance for low-aptitude perceivers while it helped improvements in performance for high-aptitude perceivers. These results show that the previously observed interaction between individuals’ aptitude and effects of degree of variability extends to natural tokens of Mandarin speech. This interaction was not found, however, in a closely matched study in which native Dutch listeners were trained on the Japanese geminate/singleton consonant contrast. This may indicate that the effectiveness of high-variability training depends not only on individuals’ aptitude in speech perception but also on the nature of the categories being acquired. PMID:25505434

  6. Advancements in robust algorithm formulation for speaker identification of whispered speech

    NASA Astrophysics Data System (ADS)

    Fan, Xing

    Whispered speech is an alternative speech production mode from neutral speech, which is used by talkers intentionally in natural conversational scenarios to protect privacy and to avoid certain content from being overheard/made public. Due to the profound differences between whispered and neutral speech in production mechanism and the absence of whispered adaptation data, the performance of speaker identification systems trained with neutral speech degrades significantly. This dissertation therefore focuses on developing a robust closed-set speaker recognition system for whispered speech by using no or limited whispered adaptation data from non-target speakers. This dissertation proposes the concept of "High''/"Low'' performance whispered data for the purpose of speaker identification. A variety of acoustic properties are identified that contribute to the quality of whispered data. An acoustic analysis is also conducted to compare the phoneme/speaker dependency of the differences between whispered and neutral data in the feature domain. The observations from those acoustic analysis are new in this area and also serve as a guidance for developing robust speaker identification systems for whispered speech. This dissertation further proposes two systems for speaker identification of whispered speech. One system focuses on front-end processing. A two-dimensional feature space is proposed to search for "Low''-quality performance based whispered utterances and separate feature mapping functions are applied to vowels and consonants respectively in order to retain the speaker's information shared between whispered and neutral speech. The other system focuses on speech-mode-independent model training. The proposed method generates pseudo whispered features from neutral features by using the statistical information contained in a whispered Universal Background model (UBM) trained from extra collected whispered data from non-target speakers. Four modeling methods are proposed for the transformation estimation in order to generate the pseudo whispered features. Both of the above two systems demonstrate a significant improvement over the baseline system on the evaluation data. This dissertation has therefore contributed to providing a scientific understanding of the differences between whispered and neutral speech as well as improved front-end processing and modeling method for speaker identification of whispered speech. Such advancements will ultimately contribute to improve the robustness of speech processing systems.

  7. Eye movements during information processing tasks: individual differences and cultural effects.

    PubMed

    Rayner, Keith; Li, Xingshan; Williams, Carrick C; Cave, Kyle R; Well, Arnold D

    2007-09-01

    The eye movements of native English speakers, native Chinese speakers, and bilingual Chinese/English speakers who were either born in China (and moved to the US at an early age) or in the US were recorded during six tasks: (1) reading, (2) face processing, (3) scene perception, (4) visual search, (5) counting Chinese characters in a passage of text, and (6) visual search for Chinese characters. Across the different groups, there was a strong tendency for consistency in eye movement behavior; if fixation durations of a given viewer were long on one task, they tended to be long on other tasks (and the same tended to be true for saccade size). Some tasks, notably reading, did not conform to this pattern. Furthermore, experience with a given writing system had a large impact on fixation durations and saccade lengths. With respect to cultural differences, there was little evidence that Chinese participants spent more time looking at the background information (and, conversely less time looking at the foreground information) than the American participants. Also, Chinese participants' fixations were more numerous and of shorter duration than those of their American counterparts while viewing faces and scenes, and counting Chinese characters in text.

  8. Impact of Neighborhood Environments on Health Consciousness, Information Seeking, and Attitudes among US-Born and Non-US-Born Free Clinic Patients.

    PubMed

    Kamimura, Akiko; Ashby, Jeanie; Jess, Allison; Trinh, Ha Ngoc; Nourian, Maziar M; Finlayson, Sarah Yukie; Prudencio, Liana; Reel, Justine J

    2015-12-01

    This study examined the impact of neighborhood environments on health consciousness, information seeking, and attitudes among uninsured free clinic patients to better understand the specific needs of the population for health promotion and prevention efforts. US-born English-speaking, non-US-born English-speaking, and Spanish-speaking free clinic patients completed a self-administered survey using reliable measures in autumn 2014 (N = 769). The results of this study suggest that social cohesion is positively associated with health consciousness, information seeking, and attitudes. Lower levels of available healthy food in the community were associated with higher levels of health consciousness. Although Spanish speakers reported lower levels of the availability of healthy food, social cohesion, and access to the Internet or text messaging compared with US-born or non-US-born English speakers, they were more likely to be health conscious and have higher levels of health information seeking. Spanish speakers as well as non-US-born English speakers, were more likely to attend health education classes compared with US-born English speakers. Health education programs for free clinic patients should include strategies to increase social cohesion. Health education programs should consider the diverse needs of these individual populations to maximize the effectiveness of the programs for free clinic patients.

  9. Attentional influences on functional mapping of speech sounds in human auditory cortex

    PubMed Central

    Obleser, Jonas; Elbert, Thomas; Eulitz, Carsten

    2004-01-01

    Background The speech signal contains both information about phonological features such as place of articulation and non-phonological features such as speaker identity. These are different aspects of the 'what'-processing stream (speaker vs. speech content), and here we show that they can be further segregated as they may occur in parallel but within different neural substrates. Subjects listened to two different vowels, each spoken by two different speakers. During one block, they were asked to identify a given vowel irrespectively of the speaker (phonological categorization), while during the other block the speaker had to be identified irrespectively of the vowel (speaker categorization). Auditory evoked fields were recorded using 148-channel magnetoencephalography (MEG), and magnetic source imaging was obtained for 17 subjects. Results During phonological categorization, a vowel-dependent difference of N100m source location perpendicular to the main tonotopic gradient replicated previous findings. In speaker categorization, the relative mapping of vowels remained unchanged but sources were shifted towards more posterior and more superior locations. Conclusions These results imply that the N100m reflects the extraction of abstract invariants from the speech signal. This part of the processing is accomplished in auditory areas anterior to AI, which are part of the auditory 'what' system. This network seems to include spatially separable modules for identifying the phonological information and for associating it with a particular speaker that are activated in synchrony but within different regions, suggesting that the 'what' processing can be more adequately modeled by a stream of parallel stages. The relative activation of the parallel processing stages can be modulated by attentional or task demands. PMID:15268765

  10. Pragma- und textilinguistische Untersuchungsmethoden zur indirekten Rede im Englischen (Pragmatic and Text-Linguistic Research Methods on Indirect Discourse in English)

    ERIC Educational Resources Information Center

    Ziegesar, Detlef von

    1976-01-01

    Examines the mechanisms involved in changing from direct to indirect discourse, taking as a starting point research findings in linguistic pragmatics, which involve the practical application of speech symbol use between speaker and hearer. (Text is in German.) (IFS/WGA)

  11. An Introduction to Spoken Bolivian Quechua.

    ERIC Educational Resources Information Center

    Bills, Garland D.; And Others

    This text is intended for use in a college-level course in Bolivian Quechua for speakers of English. It is divided into thirty units and is specifically designed for a two-semester academic year of fifteen weeks a semester. The entire text is highly structured, carefully graded, and organized on the principles of the audiolingual approach to…

  12. Relationship between central auditory processing and reading skills: preliminary observations in Hebrew speaking children.

    PubMed

    Cohen-Mimran, Ravit; Sapir, Shimon

    2008-01-01

    To assess the relationships between central auditory processing (CAP) of sinusoidally modulated speech-like and non-speech acoustic signals and reading skills in shallow (pointed) and deep (unpointed) Hebrew orthographies. Twenty unselected fifth-grade Hebrew speakers performed a rate change detection (RCD) task using the aforementioned acoustic signals. They also performed reading and general ability (IQ) tests. After controlling for general ability, RCD tasks contributed a significant unique variance to the decoding skills. In addition, there was a fairly strong correlation between the score on the RCD with the speech-like stimuli and the unpointed text reading score. CAP abilities may affect reading skills, depending on the nature of orthography (deep vs shallow), at least in the Hebrew language.

  13. The Wildcat Corpus of Native- and Foreign-Accented English: Communicative Efficiency across Conversational Dyads with Varying Language Alignment Profiles

    PubMed Central

    Van Engen, Kristin J.; Baese-Berk, Melissa; Baker, Rachel E.; Choi, Arim; Kim, Midam; Bradlow, Ann R.

    2012-01-01

    This paper describes the development of the Wildcat Corpus of native- and foreign-accented English, a corpus containing scripted and spontaneous speech recordings from 24 native speakers of American English and 52 non-native speakers of English. The core element of this corpus is a set of spontaneous speech recordings, for which a new method of eliciting dialogue-based, laboratory-quality speech recordings was developed (the Diapix task). Dialogues between two native speakers of English, between two non-native speakers of English (with either shared or different L1s), and between one native and one non-native speaker of English are included and analyzed in terms of general measures of communicative efficiency. The overall finding was that pairs of native talkers were most efficient, followed by mixed native/non-native pairs and non-native pairs with shared L1. Non-native pairs with different L1s were least efficient. These results support the hypothesis that successful speech communication depends both on the alignment of talkers to the target language and on the alignment of talkers to one another in terms of native language background. PMID:21313992

  14. Articulatory settings of French-English bilingual speakers

    NASA Astrophysics Data System (ADS)

    Wilson, Ian

    2005-04-01

    The idea of a language-specific articulatory setting (AS), an underlying posture of the articulators during speech, has existed for centuries [Laver, Historiogr. Ling. 5 (1978)], but until recently it had eluded direct measurement. In an analysis of x-ray movies of French and English monolingual speakers, Gick et al. [Phonetica (in press)] link AS to inter-speech posture, allowing measurement of AS without interference from segmental targets during speech, and they give quantitative evidence showing AS to be language-specific. In the present study, ultrasound and Optotrak are used to investigate whether bilingual English-French speakers have two ASs, and whether this varies depending on the mode (monolingual or bilingual) these speakers are in. Specifically, for inter-speech posture of the lips, lip aperture and protrusion are measured using Optotrak. For inter-speech posture of the tongue, tongue root retraction, tongue body and tongue tip height are measured using optically-corrected ultrasound. Segmental context is balanced across the two languages ensuring that the sets of sounds before and after an inter-speech posture are consistent across languages. By testing bilingual speakers, vocal tract morphology across languages is controlled for. Results have implications for L2 acquisition, specifically the teaching and acquisition of pronunciation.

  15. On compensation of mismatched recording conditions in the Bayesian approach for forensic automatic speaker recognition.

    PubMed

    Botti, F; Alexander, A; Drygajlo, A

    2004-12-02

    This paper deals with a procedure to compensate for mismatched recording conditions in forensic speaker recognition, using a statistical score normalization. Bayesian interpretation of the evidence in forensic automatic speaker recognition depends on three sets of recordings in order to perform forensic casework: reference (R) and control (C) recordings of the suspect, and a potential population database (P), as well as a questioned recording (QR) . The requirement of similar recording conditions between suspect control database (C) and the questioned recording (QR) is often not satisfied in real forensic cases. The aim of this paper is to investigate a procedure of normalization of scores, which is based on an adaptation of the Test-normalization (T-norm) [2] technique used in the speaker verification domain, to compensate for the mismatch. Polyphone IPSC-02 database and ASPIC (an automatic speaker recognition system developed by EPFL and IPS-UNIL in Lausanne, Switzerland) were used in order to test the normalization procedure. Experimental results for three different recording condition scenarios are presented using Tippett plots and the effect of the compensation on the evaluation of the strength of the evidence is discussed.

  16. Discourse Factors Influencing Spatial Descriptions in English and German

    NASA Astrophysics Data System (ADS)

    Vorwerg, Constanze; Tenbrink, Thora

    The ways in which objects are referred to by using spatial language depend on many factors, including the spatial configuration and the discourse context. We present the results of a web experiment in which speakers were asked to either describe where a specified item was located in a picture containing several items, or which item was specified. Furthermore, conditions differed as to whether the first six configurations were specifically simple or specifically complex. Results show that speakers' spatial descriptions are more detailed if the question is where rather than which, mirroring the fact that contrasting the target item from the others in which tasks may not always require an equally detailed spatial description as in where tasks. Furthermore, speakers are influenced by the complexity of initial configurations in intricate ways: on the one hand, individual speakers tend to self-align with respect to their earlier linguistic strategies; however, also a contrast effect could be identified with respect to the usage of combined projective terms.

  17. The effect of simultaneous text on the recall of noise-degraded speech.

    PubMed

    Grossman, Irina; Rajan, Ramesh

    2017-05-01

    Written and spoken language utilize the same processing system, enabling text to modulate speech processing. We investigated how simultaneously presented text affected speech recall in babble noise using a retrospective recall task. Participants were presented with text-speech sentence pairs in multitalker babble noise and then prompted to recall what they heard or what they read. In Experiment 1, sentence pairs were either congruent or incongruent and they were presented in silence or at 1 of 4 noise levels. Audio and Visual control groups were also tested with sentences presented in only 1 modality. Congruent text facilitated accurate recall of degraded speech; incongruent text had no effect. Text and speech were seldom confused for each other. A consideration of the effects of the language background found that monolingual English speakers outperformed early multilinguals at recalling degraded speech; however the effects of text on speech processing were analogous. Experiment 2 considered if the benefit provided by matching text was maintained when the congruency of the text and speech becomes more ambiguous because of the addition of partially mismatching text-speech sentence pairs that differed only on their final keyword and because of the use of low signal-to-noise ratios. The experiment focused on monolingual English speakers; the results showed that even though participants commonly confused text-for-speech during incongruent text-speech pairings, these confusions could not fully account for the benefit provided by matching text. Thus, we uniquely demonstrate that congruent text benefits the recall of noise-degraded speech. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  18. Robust speaker's location detection in a vehicle environment using GMM models.

    PubMed

    Hu, Jwu-Sheng; Cheng, Chieh-Cheng; Liu, Wei-Han

    2006-04-01

    Abstract-Human-computer interaction (HCI) using speech communication is becoming increasingly important, especially in driving where safety is the primary concern. Knowing the speaker's location (i.e., speaker localization) not only improves the enhancement results of a corrupted signal, but also provides assistance to speaker identification. Since conventional speech localization algorithms suffer from the uncertainties of environmental complexity and noise, as well as from the microphone mismatch problem, they are frequently not robust in practice. Without a high reliability, the acceptance of speech-based HCI would never be realized. This work presents a novel speaker's location detection method and demonstrates high accuracy within a vehicle cabinet using a single linear microphone array. The proposed approach utilize Gaussian mixture models (GMM) to model the distributions of the phase differences among the microphones caused by the complex characteristic of room acoustic and microphone mismatch. The model can be applied both in near-field and far-field situations in a noisy environment. The individual Gaussian component of a GMM represents some general location-dependent but content and speaker-independent phase difference distributions. Moreover, the scheme performs well not only in nonline-of-sight cases, but also when the speakers are aligned toward the microphone array but at difference distances from it. This strong performance can be achieved by exploiting the fact that the phase difference distributions at different locations are distinguishable in the environment of a car. The experimental results also show that the proposed method outperforms the conventional multiple signal classification method (MUSIC) technique at various SNRs.

  19. Perceptions of Patient-Provider Communication in Breast and Cervical Cancer-Related Care: A Qualitative Study of Low-Income English- and Spanish-Speaking Women

    PubMed Central

    Simon, Melissa A.; Ragas, Daiva M.; Nonzee, Narissa J.; Phisuthikul, Ava M.; Luu, Thanh Ha; Dong, XinQi

    2013-01-01

    To explore patient perceptions of patient-provider communication in breast and cervical cancer-related care among low-income English- and Spanish- speaking women, we examined communication barriers and facilitators reported by patients receiving care at safety net clinics. Participants were interviewed in English or Spanish after receiving an abnormal breast or cervical cancer screening test or cancer diagnosis. Following an inductive approach, interviews were coded and analyzed by the language spoken with providers and patient-provider language concordance status. Of 78 participants, 53% (n = 41) were English-speakers and 47% (n = 37) were Spanish-speakers. All English-speakers were language-concordant with providers. Of Spanish-speakers, 27% (n = 10) were Spanish-concordant; 38% (n = 14) were Spanish-discordant, requiring an interpreter; and 35% (n = 13) were Spanish mixed-concordant, experiencing both types of communication throughout the care continuum. English-speakers focused on communication barriers, and difficulty understanding jargon arose as a theme. Spanish-speakers emphasized communication facilitators related to Spanish language use. Themes among all Spanish-speaking sub-groups included appreciation for language support resources and preference for Spanish-speaking providers. Mixed-concordant participants accounted for the majority of Spanish-speakers who reported communication barriers. Our data suggest that, although perception of patient-provider communication may depend on the language spoken throughout the care continuum, jargon is lost when health information is communicated in Spanish. Further, the respective consistency of language concordance or interpretation may play a role in patient perception of patient-provider communication. PMID:23553683

  20. Exploring the Lack of a Disfluency Effect: Evidence from Eye Movements

    ERIC Educational Resources Information Center

    Strukelj, Alexander; Scheiter, Katharina; Nyström, Marcus; Holmqvist, Kenneth

    2016-01-01

    An eye-tracking study with 60 native Swedish speakers (18-30 years) was conducted to investigate the positive effects on learning outcomes predicted by the disfluency effect. Subtle low-pass filtering was used as a disfluency manipulation and compared with a control condition using regular text. The text was presented on four separate text…

  1. Shared Features of L2 Writing: Intergroup Homogeneity and Text Classification

    ERIC Educational Resources Information Center

    Crossley, Scott A.; McNamara, Danielle S.

    2011-01-01

    This study investigates intergroup homogeneity within high intermediate and advanced L2 writers of English from Czech, Finnish, German, and Spanish first language backgrounds. A variety of linguistic features related to lexical sophistication, syntactic complexity, and cohesion were used to compare texts written by L1 speakers of English to L2…

  2. The Effect of Bilingualism on Communication Efficiency in Text Messages (SMS)

    ERIC Educational Resources Information Center

    Carrier, L. Mark; Benitez, Sandra Y.

    2010-01-01

    The widespread use of cell phones has led to the proliferation of messages sent using the Short Messaging Service (SMS). The 160-character limit on text messages encourages the use of shortenings and other shortcuts in language use. When bilingual speakers use SMS, their access to multiple sources of vocabulary, sentence structure, and other…

  3. "Siamo Vicini, No?": Negotiating Commonality for Rapport Building in Italian L1-L2 Online Text Chat

    ERIC Educational Resources Information Center

    Tudini, Vincenza; Strambi, Antonella

    2017-01-01

    Focusing specifically on the negotiation of commonality, this study explores rapport building (Spencer-Oatey, 2000) in online intercultural text chat, where Australian students of Italian interact with L1 Italian speakers. Although the initial purpose of the examined chat exchanges is to facilitate L2 acquisition, analysis of transcripts indicates…

  4. Las Expresiones Pragmaticas en la Conversacion Espontanea (Pragmatic Expressions in Spontaneous Conversation).

    ERIC Educational Resources Information Center

    Carranza, Isolda

    The starting point of this thesis is the hypothesis that in Spanish there are conventionalized expressions that signal both the articulation of text parts and the speaker's attitude towards the utterance. "Pragmatic Expressions" (PEs) are fixed lexical forms that contribute to text creation in relation to the context of enunciation, are oriented…

  5. Text-Based Negotiated Interaction of NNS-NNS and NNS-NS Dyads on Facebook

    ERIC Educational Resources Information Center

    Liu, Sarah Hsueh-Jui

    2017-01-01

    This study sought to determine the difference in text-based negotiated interaction between non-native speakers of English (NNS-NNS) and between non-native and natives (NNS-NS) in terms of the frequency of negotiated instances, successfully resolved instances, and interactional strategy use when the dyads collaborated on Facebook. It involved 10…

  6. A Grammar of Spoken Brazilian Portuguese.

    ERIC Educational Resources Information Center

    Thomas, Earl W.

    This is a first-year text of Portuguese grammar based on the Portuguese of moderately educated Brazilians from the area around Rio de Janeiro. Spoken idiomatic usage is emphasized. An important innovation is found in the presentation of verb tenses; they are presented in the order in which the native speaker learns them. The text is intended to…

  7. GMM-based speaker age and gender classification in Czech and Slovak

    NASA Astrophysics Data System (ADS)

    Přibil, Jiří; Přibilová, Anna; Matoušek, Jindřich

    2017-01-01

    The paper describes an experiment with using the Gaussian mixture models (GMM) for automatic classification of the speaker age and gender. It analyses and compares the influence of different number of mixtures and different types of speech features used for GMM gender/age classification. Dependence of the computational complexity on the number of used mixtures is also analysed. Finally, the GMM classification accuracy is compared with the output of the conventional listening tests. The results of these objective and subjective evaluations are in correspondence.

  8. The Directed Megaphone: A Theater Commander’s Means to Communicate His Vision and Intent

    DTIC Science & Technology

    1993-05-01

    commander must have oratory skills--both logos and pathos , as described by Aristotle . He must have a flair for the dramatic to embellish his message...went on to say that persuasion depends on three elements: logos -- The truth and logical validity of what is being argued. ethos -- The speakers...presents the message ( pathos ), and who the speaker is (ethoas). Many experts in the field of communications identify ethos as the most important persuasive

  9. The dynamic effect of reading direction habit on spatial asymmetry of image perception.

    PubMed

    Afsari, Zaeinab; Ossandón, José P; König, Peter

    2016-09-01

    Exploration of images after stimulus onset is initially biased to the left. Here, we studied the causes of such an asymmetry and investigated effects of reading habits, text primes, and priming by systematically biased eye movements on this spatial bias in visual exploration. Bilinguals first read text primes with right-to-left (RTL) or left-to-right (LTR) reading directions and subsequently explored natural images. In Experiment 1, native RTL speakers showed a leftward free-viewing shift after reading LTR primes but a weaker rightward bias after reading RTL primes. This demonstrates that reading direction dynamically influences the spatial bias. However, native LTR speakers who learned an RTL language late in life showed a leftward bias after reading either LTR or RTL primes, which suggests the role of habit formation in the production of the spatial bias. In Experiment 2, LTR bilinguals showed a slightly enhanced leftward bias after reading LTR text primes in their second language. This might contribute to the differences of native RTL and LTR speakers observed in Experiment 1. In Experiment 3, LTR bilinguals read normal (LTR, habitual reading) and mirrored left-to-right (mLTR, nonhabitual reading) texts. We observed a strong leftward bias in both cases, indicating that the bias direction is influenced by habitual reading direction and is not secondary to the actual reading direction. This is confirmed in Experiment 4, in which LTR participants were asked to follow RTL and LTR moving dots in prior image presentation and showed no change in the normal spatial bias. In conclusion, the horizontal bias is a dynamic property and is modulated by habitual reading direction.

  10. Hyper-active gap filling

    PubMed Central

    Omaki, Akira; Lau, Ellen F.; Davidson White, Imogen; Dakan, Myles L.; Apple, Aaron; Phillips, Colin

    2015-01-01

    Much work has demonstrated that speakers of verb-final languages are able to construct rich syntactic representations in advance of verb information. This may reflect general architectural properties of the language processor, or it may only reflect a language-specific adaptation to the demands of verb-finality. The present study addresses this issue by examining whether speakers of a verb-medial language (English) wait to consult verb transitivity information before constructing filler-gap dependencies, where internal arguments are fronted and hence precede the verb. This configuration makes it possible to investigate whether the parser actively makes representational commitments on the gap position before verb transitivity information becomes available. A key prediction of the view that rich pre-verbal structure building is a general architectural property is that speakers of verb-medial languages should predictively construct dependencies in advance of verb transitivity information, and therefore that disruption should be observed when the verb has intransitive subcategorization frames that are incompatible with the predicted structure. In three reading experiments (self-paced and eye-tracking) that manipulated verb transitivity, we found evidence for reading disruption when the verb was intransitive, although no such reading difficulty was observed when the critical verb was embedded inside a syntactic island structure, which blocks filler-gap dependency completion. These results are consistent with the hypothesis that in English, as in verb-final languages, information from preverbal noun phrases is sufficient to trigger active dependency completion without having access to verb transitivity information. PMID:25914658

  11. Hyper-active gap filling.

    PubMed

    Omaki, Akira; Lau, Ellen F; Davidson White, Imogen; Dakan, Myles L; Apple, Aaron; Phillips, Colin

    2015-01-01

    Much work has demonstrated that speakers of verb-final languages are able to construct rich syntactic representations in advance of verb information. This may reflect general architectural properties of the language processor, or it may only reflect a language-specific adaptation to the demands of verb-finality. The present study addresses this issue by examining whether speakers of a verb-medial language (English) wait to consult verb transitivity information before constructing filler-gap dependencies, where internal arguments are fronted and hence precede the verb. This configuration makes it possible to investigate whether the parser actively makes representational commitments on the gap position before verb transitivity information becomes available. A key prediction of the view that rich pre-verbal structure building is a general architectural property is that speakers of verb-medial languages should predictively construct dependencies in advance of verb transitivity information, and therefore that disruption should be observed when the verb has intransitive subcategorization frames that are incompatible with the predicted structure. In three reading experiments (self-paced and eye-tracking) that manipulated verb transitivity, we found evidence for reading disruption when the verb was intransitive, although no such reading difficulty was observed when the critical verb was embedded inside a syntactic island structure, which blocks filler-gap dependency completion. These results are consistent with the hypothesis that in English, as in verb-final languages, information from preverbal noun phrases is sufficient to trigger active dependency completion without having access to verb transitivity information.

  12. 75 FR 9182 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-03-01

    ... Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification to fulfill the... the Speaker of the House of Representatives, Transmittals 09-28 with attached transmittal, and policy...

  13. Sound Processing Features for Speaker-Dependent and Phrase-Independent Emotion Recognition in Berlin Database

    NASA Astrophysics Data System (ADS)

    Anagnostopoulos, Christos Nikolaos; Vovoli, Eftichia

    An emotion recognition framework based on sound processing could improve services in human-computer interaction. Various quantitative speech features obtained from sound processing of acting speech were tested, as to whether they are sufficient or not to discriminate between seven emotions. Multilayered perceptrons were trained to classify gender and emotions on the basis of a 24-input vector, which provide information about the prosody of the speaker over the entire sentence using statistics of sound features. Several experiments were performed and the results were presented analytically. Emotion recognition was successful when speakers and utterances were “known” to the classifier. However, severe misclassifications occurred during the utterance-independent framework. At least, the proposed feature vector achieved promising results for utterance-independent recognition of high- and low-arousal emotions.

  14. Functional activity and white matter microstructure reveal the independent effects of age of acquisition and proficiency on second-language learning.

    PubMed

    Nichols, Emily S; Joanisse, Marc F

    2016-12-01

    Two key factors govern how bilingual speakers neurally maintain two languages: the speakers' second language age of acquisition (AoA) and their subsequent proficiency. However, the relative roles of these two factors have been difficult to disentangle given that the two can be closely correlated, and most prior studies have examined the two factors in isolation. Here, we combine functional magnetic resonance imaging with diffusion tensor imaging to identify specific brain areas that are independently modulated by AoA and proficiency in second language speakers. First-language Mandarin Chinese speakers who are second language speakers of English were scanned as they performed a picture-word matching task in either language. In the same session we also acquired diffusion-weighted scans to assess white matter microstructure, along with behavioural measures of language proficiency prior to entering the scanner. Results reveal gray- and white-matter networks involving both the left and right hemisphere that independently vary as a function of a second-language speaker's AoA and proficiency, focused on the superior temporal gyrus, middle and inferior frontal gyrus, parahippocampal gyrus, and the basal ganglia. These results indicate that proficiency and AoA explain separate functional and structural networks in the bilingual brain, which we interpret as suggesting distinct types of plasticity for age-dependent effects (i.e., AoA) versus experience and/or predisposition (i.e., proficiency). Copyright © 2016 The Authors. Published by Elsevier Inc. All rights reserved.

  15. Fifty years of progress in speech and speaker recognition

    NASA Astrophysics Data System (ADS)

    Furui, Sadaoki

    2004-10-01

    Speech and speaker recognition technology has made very significant progress in the past 50 years. The progress can be summarized by the following changes: (1) from template matching to corpus-base statistical modeling, e.g., HMM and n-grams, (2) from filter bank/spectral resonance to Cepstral features (Cepstrum + DCepstrum + DDCepstrum), (3) from heuristic time-normalization to DTW/DP matching, (4) from gdistanceh-based to likelihood-based methods, (5) from maximum likelihood to discriminative approach, e.g., MCE/GPD and MMI, (6) from isolated word to continuous speech recognition, (7) from small vocabulary to large vocabulary recognition, (8) from context-independent units to context-dependent units for recognition, (9) from clean speech to noisy/telephone speech recognition, (10) from single speaker to speaker-independent/adaptive recognition, (11) from monologue to dialogue/conversation recognition, (12) from read speech to spontaneous speech recognition, (13) from recognition to understanding, (14) from single-modality (audio signal only) to multi-modal (audio/visual) speech recognition, (15) from hardware recognizer to software recognizer, and (16) from no commercial application to many practical commercial applications. Most of these advances have taken place in both the fields of speech recognition and speaker recognition. The majority of technological changes have been directed toward the purpose of increasing robustness of recognition, including many other additional important techniques not noted above.

  16. Perception of a non-native speech contrast: Voiced and voiceless stops as perceived by Tamil speakers

    NASA Astrophysics Data System (ADS)

    Tur, Sylwia

    2004-05-01

    The effect of linguistic experience plays a significant role in how speech sounds are perceived. The findings of many studies imply that the perception of non-native contrasts depends on their status in the native language of the listener. Tamil is a language with a single voicing category. All stop consonants in Tamil are phonemically voiceless, though allophonic voicing has been observed in spoken Tamil. The present study examined how native Tamil speakers and English controls perceived voiced and voiceless bilabial, alveolar, and velar stops in English. Voice onset time (VOT) was manipulated for editing of naturally produced stimuli with increasingly longer continuum. Perceptual data was collected from 16 Tamil and 16 English speakers. Experiment 1 was an AX task in which subjects responded same or different to 162 pairs of stimuli. Experiment 2 was a forced choice ID task in which subjects identified 99 individually presented stimuli as pa, ta, ka or ba, da, ga. Experiments show statistically significant differences between Tamil and English speakers in their perception of English stop consonants. Results of the study imply that the allophonic status of voiced stops in Tamil does not aid the Tamil speakers in perceiving phonemically voiced stops in English.

  17. A Method for Determining the Timing of Displaying the Speaker's Face and Captions for a Real-Time Speech-to-Caption System

    NASA Astrophysics Data System (ADS)

    Kuroki, Hayato; Ino, Shuichi; Nakano, Satoko; Hori, Kotaro; Ifukube, Tohru

    The authors of this paper have been studying a real-time speech-to-caption system using speech recognition technology with a “repeat-speaking” method. In this system, they used a “repeat-speaker” who listens to a lecturer's voice and then speaks back the lecturer's speech utterances into a speech recognition computer. The througoing system showed that the accuracy of the captions is about 97% in Japanese-Japanese conversion and the conversion time from voices to captions is about 4 seconds in English-English conversion in some international conferences. Of course it required a lot of costs to achieve these high performances. In human communications, speech understanding depends not only on verbal information but also on non-verbal information such as speaker's gestures, and face and mouth movements. So the authors found the idea to display information of captions and speaker's face movement images with a suitable way to achieve a higher comprehension after storing information once into a computer briefly. In this paper, we investigate the relationship of the display sequence and display timing between captions that have speech recognition errors and the speaker's face movement images. The results show that the sequence “to display the caption before the speaker's face image” improves the comprehension of the captions. The sequence “to display both simultaneously” shows an improvement only a few percent higher than the question sentence, and the sequence “to display the speaker's face image before the caption” shows almost no change. In addition, the sequence “to display the caption 1 second before the speaker's face shows the most significant improvement of all the conditions.

  18. Linea abierta -- Comunicacion en espanol. Nivel Uno (Open Line -- Communication in Spanish. Level One).

    ERIC Educational Resources Information Center

    Dissemination and Assessment Center for Bilingual Education, Austin, TX.

    This text is the first of a series of four units concerning language skills for Spanish speakers of limited English-speaking ability in grades 6-9. The text includes a variety of learning experiences designed to develop competence in the student's cognitive and affective skills derived from the following basic areas: listening, speaking, reading,…

  19. Linea abierta -- Comunicacion en Espanol. Edicion del maestro (Open Line -- Communication in Spanish. Teacher's Edition).

    ERIC Educational Resources Information Center

    Dissemination and Assessment Center for Bilingual Education, Austin, TX.

    This teacher's guide corresponds to the text that is the first of a series of four units concerning language skills for Spanish speakers of limited English-speaking ability in grades 6-9. The text includes a variety of learning experiences designed to develop competence in the student's cognitive and affective skills derived from the following…

  20. Using Short Texts to Teach English as Second Language: An Integrated Approach

    ERIC Educational Resources Information Center

    Kembo, Jane

    2016-01-01

    The teacher of English Language is often hard pressed to find interesting and authentic ways to present language to target second language speakers. While language can be taught and learned, part of it must be acquired and short texts provide powerful tools for doing so and reinforcing what has been taught/learned. This paper starts from research,…

  1. The Role of Vowels in Reading Semitic Scripts: Data from Arabic and Hebrew.

    ERIC Educational Resources Information Center

    Abu-Rabia, Salim

    2001-01-01

    Investigates the effect of vowels and context on reading accuracy of skilled adult native Arabic speakers in Arabic and in Hebrew, their second language. Reveals a significant effect for vowels and for context across all reading conditions in Arabic and Hebrew. Finds that the vowelized texts in Arabic and the pointed and unpointed texts in Hebrew…

  2. 78 FR 69124 - Trinity Adaptive Management Working Group; Public Meeting and Teleconference

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-11-18

    ... Recommendation, Bylaw discussion, 2014 Flow Alternatives, Status of Klamath fall flow release, Mining issues...Point, or rich text file). Registered speakers who wish to expand on their oral statements, or those who...

  3. Sociolinguistic and psycholinguistic indications of behavior disorders: analysis of a prisoner's discourse.

    PubMed

    Timor, Uri; Weiss, Joshua M

    2008-02-01

    Human verbal language communicates both manifest and latent messages concerning the speaker's world and behavior. To understand his world and analyze his problems,(1) it is important to decode the latent messages as they may hint at the root causes. The authors present a discourse analysis of a prisoner's text and a semantic and morphological analysis of it. This text reflects contempt for the law and its representatives, together with a weak attachment to legitimate society, neutralization of personal responsibility, denial of guilt, and low self-esteem. Sociolinguistic and psycholinguistic analysis points toward a more profound evaluation of the perceptions and world of the speaker. It seems that he yearns for attachment, for understanding and social acceptance, and perhaps even to abandon crime. The prisoner's latent feelings of helplessness and fear of humiliation may help the therapist establish a therapeutic relationship and help him change his perceptions and behavior.

  4. Developing a corpus of spoken language variability

    NASA Astrophysics Data System (ADS)

    Carmichael, Lesley; Wright, Richard; Wassink, Alicia Beckford

    2003-10-01

    We are developing a novel, searchable corpus as a research tool for investigating phonetic and phonological phenomena across various speech styles. Five speech styles have been well studied independently in previous work: reduced (casual), careful (hyperarticulated), citation (reading), Lombard effect (speech in noise), and ``motherese'' (child-directed speech). Few studies to date have collected a wide range of styles from a single set of speakers, and fewer yet have provided publicly available corpora. The pilot corpus includes recordings of (1) a set of speakers participating in a variety of tasks designed to elicit the five speech styles, and (2) casual peer conversations and wordlists to illustrate regional vowels. The data include high-quality recordings and time-aligned transcriptions linked to text files that can be queried. Initial measures drawn from the database provide comparison across speech styles along the following acoustic dimensions: MLU (changes in unit duration); relative intra-speaker intensity changes (mean and dynamic range); and intra-speaker pitch values (minimum, maximum, mean, range). The corpus design will allow for a variety of analyses requiring control of demographic and style factors, including hyperarticulation variety, disfluencies, intonation, discourse analysis, and detailed spectral measures.

  5. Gender parity trends for invited speakers at four prominent virology conference series.

    PubMed

    Kalejta, Robert F; Palmenberg, Ann C

    2017-06-07

    Scientific conferences are most beneficial to participants when they showcase significant new experimental developments, accurately summarize the current state of the field, and provide strong opportunities for collaborative networking. A top-notch slate of invited speakers, assembled by conference organizers or committees, is key to achieving these goals. The perceived underrepresentation of female speakers at prominent scientific meetings is currently a popular topic for discussion, but one that often lacks supportive data. We compiled the full rosters of invited speakers over the last 35 years for four prominent international virology conferences, the American Society for Virology Annual Meeting (ASV), the International Herpesvirus Workshop (IHW), the Positive-Strand RNA Virus Symposium (PSR), and the Gordon Research Conference on Viruses & Cells (GRC). The rosters were cross-indexed by unique names, gender, year, and repeat invitations. When plotted as gender-dependent trends over time, all four conferences showed a clear proclivity for male-dominated invited speaker lists. Encouragingly, shifts toward parity are emerging within all units, but at different rates. Not surprisingly, both selection of a larger percentage of first time participants and the presence of a woman on the speaker selection committee correlated with improved parity. Session chair information was also collected for the IHW and GRC. These visible positions also displayed a strong male dominance over time that is eroding slowly. We offer our personal interpretation of these data to aid future organizers achieve improved equity among the limited number of available positions for session moderators and invited speakers. IMPORTANCE Politicians and media members have a tendency to cite anecdotes as conclusions without any supporting data. This happens so frequently now, that a name for it has emerged: fake news. Good science proceeds otherwise. The under representation of women as invited speakers at international scientific conferences exemplifies a present-day discussion topic usually occurring without facts to support or refute the arguments. We now provide records profiling four prominent virology conferences over the years 1982 to 2017 with the intention that the trends and accompanying analyses of the gender parity of invited speakers may allow the ongoing discussions to be informed. Copyright © 2017 American Society for Microbiology.

  6. Gender Parity Trends for Invited Speakers at Four Prominent Virology Conference Series

    PubMed Central

    Palmenberg, Ann C.

    2017-01-01

    ABSTRACT Scientific conferences are most beneficial to participants when they showcase significant new experimental developments, accurately summarize the current state of the field, and provide strong opportunities for collaborative networking. A top-notch slate of invited speakers, assembled by conference organizers or committees, is key to achieving these goals. The perceived underrepresentation of female speakers at prominent scientific meetings is currently a popular topic for discussion, but one that often lacks supportive data. We compiled the full rosters of invited speakers over the last 35 years for four prominent international virology conferences, the American Society for Virology Annual Meeting (ASV), the International Herpesvirus Workshop (IHW), the Positive-Strand RNA Virus Symposium (PSR), and the Gordon Research Conference on Viruses & Cells (GRC). The rosters were cross-indexed by unique names, gender, year, and repeat invitations. When plotted as gender-dependent trends over time, all four conferences showed a clear proclivity for male-dominated invited speaker lists. Encouragingly, shifts toward parity are emerging within all units, but at different rates. Not surprisingly, both selection of a larger percentage of first-time participants and the presence of a woman on the speaker selection committee correlated with improved parity. Session chair information was also collected for the IHW and GRC. These visible positions also displayed a strong male dominance over time that is eroding slowly. We offer our personal interpretation of these data to aid future organizers achieve improved equity among the limited number of available positions for session moderators and invited speakers. IMPORTANCE Politicians and media members have a tendency to cite anecdotes as conclusions without any supporting data. This happens so frequently now, that a name for it has emerged: fake news. Good science proceeds otherwise. The underrepresentation of women as invited speakers at international scientific conferences exemplifies a present-day discussion topic usually occurring without facts to support or refute the arguments. We now provide records profiling four prominent virology conferences over the years 1982 to 2017 with the intention that the trends and accompanying analyses of the gender parity of invited speakers may allow the ongoing discussions to be informed. PMID:28592542

  7. Variability in the relationships among voice quality, harmonic amplitudes, open quotient, and glottal area waveform shape in sustained phonationa

    PubMed Central

    Kreiman, Jody; Shue, Yen-Liang; Chen, Gang; Iseli, Markus; Gerratt, Bruce R.; Neubauer, Juergen; Alwan, Abeer

    2012-01-01

    Increases in open quotient are widely assumed to cause changes in the amplitude of the first harmonic relative to the second (H1*–H2*), which in turn correspond to increases in perceived vocal breathiness. Empirical support for these assumptions is rather limited, and reported relationships among these three descriptive levels have been variable. This study examined the empirical relationship among H1*–H2*, the glottal open quotient (OQ), and glottal area waveform skewness, measured synchronously from audio recordings and high-speed video images of the larynges of six phonetically knowledgeable, vocally healthy speakers who varied fundamental frequency and voice qualities quasi-orthogonally. Across speakers and voice qualities, OQ, the asymmetry coefficient, and fundamental frequency accounted for an average of 74% of the variance in H1*–H2*. However, analyses of individual speakers showed large differences in the strategies used to produce the same intended voice qualities. Thus, H1*–H2* can be predicted with good overall accuracy, but its relationship to phonatory characteristics appears to be speaker dependent. PMID:23039455

  8. Capturing patient information at nursing shift changes: methodological evaluation of speech recognition and information extraction

    PubMed Central

    Suominen, Hanna; Johnson, Maree; Zhou, Liyuan; Sanchez, Paula; Sirel, Raul; Basilakis, Jim; Hanlen, Leif; Estival, Dominique; Dawson, Linda; Kelly, Barbara

    2015-01-01

    Objective We study the use of speech recognition and information extraction to generate drafts of Australian nursing-handover documents. Methods Speech recognition correctness and clinicians’ preferences were evaluated using 15 recorder–microphone combinations, six documents, three speakers, Dragon Medical 11, and five survey/interview participants. Information extraction correctness evaluation used 260 documents, six-class classification for each word, two annotators, and the CRF++ conditional random field toolkit. Results A noise-cancelling lapel-microphone with a digital voice recorder gave the best correctness (79%). This microphone was also the most preferred option by all but one participant. Although the participants liked the small size of this recorder, their preference was for tablets that can also be used for document proofing and sign-off, among other tasks. Accented speech was harder to recognize than native language and a male speaker was detected better than a female speaker. Information extraction was excellent in filtering out irrelevant text (85% F1) and identifying text relevant to two classes (87% and 70% F1). Similarly to the annotators’ disagreements, there was confusion between the remaining three classes, which explains the modest 62% macro-averaged F1. Discussion We present evidence for the feasibility of speech recognition and information extraction to support clinicians’ in entering text and unlock its content for computerized decision-making and surveillance in healthcare. Conclusions The benefits of this automation include storing all information; making the drafts available and accessible almost instantly to everyone with authorized access; and avoiding information loss, delays, and misinterpretations inherent to using a ward clerk or transcription services. PMID:25336589

  9. Connected word recognition using a cascaded neuro-computational model

    NASA Astrophysics Data System (ADS)

    Hoya, Tetsuya; van Leeuwen, Cees

    2016-10-01

    We propose a novel framework for processing a continuous speech stream that contains a varying number of words, as well as non-speech periods. Speech samples are segmented into word-tokens and non-speech periods. An augmented version of an earlier-proposed, cascaded neuro-computational model is used for recognising individual words within the stream. Simulation studies using both a multi-speaker-dependent and speaker-independent digit string database show that the proposed method yields a recognition performance comparable to that obtained by a benchmark approach using hidden Markov models with embedded training.

  10. Developing Appreciation for Sarcasm and Sarcastic Gossip: It Depends on Perspective.

    PubMed

    Glenwright, Melanie; Tapley, Brent; Rano, Jacqueline K S; Pexman, Penny M

    2017-11-09

    Speakers use sarcasm to criticize others and to be funny; the indirectness of sarcasm protects the addressee's face (Brown & Levinson, 1987). Thus, appreciation of sarcasm depends on the ability to consider perspectives. We investigated development of this ability from late childhood into adulthood and examined effects of interpretive perspective and parties present. We presented 9- to 10-year-olds, 13- to 14-year-olds, and adults with sarcastic and literal remarks in three parties-present conditions: private evaluation, public evaluation, and gossip. Participants interpreted the speaker's attitude and humor from the addressee's perspective and, when appropriate, from the bystander's perspective. Children showed no influence of interpretive perspective or parties present on appreciation of the speaker's attitude or humor. Adolescents and adults, however, shifted their interpretations, judging that addressees have less favorable views of criticisms than bystanders. Further, adolescents and adults differed in their perceptions of the social functions of gossip, with adolescents showing more positive attitudes than adults toward sarcastic gossip. We suggest that adults' disapproval of sarcastic gossip shows a deeper understanding of the utility of sarcasm's face-saving function. Thus, the ability to modulate appreciation of sarcasm according to interpretive perspective and parties present continues to develop in adolescence and into adulthood.

  11. Effects of Dictation, Speech to Text, and Handwriting on the Written Composition of Elementary School English Language Learners

    ERIC Educational Resources Information Center

    Arcon, Nina; Klein, Perry D.; Dombroski, Jill D.

    2017-01-01

    Previous research has shown that both dictation and speech-to-text (STT) software can increase the quality of writing for native English speakers. The purpose of this study was to investigate the effect of these modalities on the written composition and cognitive load of elementary school English language learners (ELLs). In a within-subjects…

  12. Speech-Act and Text-Act Theory: "Theme-ing" in Freshman Composition.

    ERIC Educational Resources Information Center

    Horner, Winifred B.

    In contrast to a speech-act theory that is limited by a simple speaker/hearer relationship, a text-act theory of written language allows for the historical or personal context of a writer and reader, both in the written work itself and in the act of reading. This theory can be applied to theme writing, essay examinations, and revision in the…

  13. The Library Environment, Aspects of Interior Planning. Proceedings of the Library Equipment Institute (St. Louis, Missouri, June 26-27, 1964).

    ERIC Educational Resources Information Center

    Poole, Frazer G., Ed.

    The texts of the papers given at the Library Equipment Institute, as well as the presentations of the panelists and the discussions between members of the audience and the program speakers, are included. Diagrams and other illustrative materials accompany the text. Specific topics include--(1) furnishings, (2) illumination, (3) audio, (4)…

  14. Abraham Lincoln, Laurent Clerc, and the Design of the World: Lincoln Day Address at Gallaudet University, February 11, 2009

    ERIC Educational Resources Information Center

    Baynton, Douglas C.

    2010-01-01

    Protestant ministers often construct their sermons around a text from the Bible that they expand upon to make some broader point. In the nineteenth century, public speakers frequently used the same rhetorical formula, taking their text not necessarily from the Bible but from any well-known source. In his famous Cooper Union speech of 1860, Abraham…

  15. Masking Release for Igbo and English.

    PubMed

    Ebem, Deborah U; Desloge, Joseph G; Reed, Charlotte M; Braida, Louis D; Uguru, Joy O

    2013-09-01

    In this research, we explored the effect of noise interruption rate on speech intelligibility. Specifically, we used the Hearing In Noise Test (HINT) procedure with the original HINT stimuli (English) and Igbo stimuli to assess speech reception ability in interrupted noise. For a given noise level, the HINT test provides an estimate of the signal-to-noise ratio (SNR) required for 50%-correct speech intelligibility. The SNR for 50%-correct intelligibility changes depending upon the interruption rate of the noise. This phenomenon (called Masking Release) has been studied extensively in English but not for Igbo - which is an African tonal language spoken predominantly in South Eastern Nigeria. This experiment explored and compared the phenomenon of Masking Release for (i) native English speakers listening to English, (ii) native Igbo speakers listening to English, and (iii) native Igbo speakers listening to Igbo. Since Igbo is a tonal language and English is a non-tonal language, this allowed us to compare Masking Release patterns on native speakers of tonal and non-tonal languages. Our results for native English speakers listening to English HINT show that the SNR and the masking release are orderly and consistent with other English HINT data for English speakers. Our result for Igbo speakers listening to English HINT sentences show that there is greater variability in results across the different Igbo listeners than across the English listeners. This result likely reflects different levels of ability in the English language across the Igbo listeners. The masking release values in dB are less than for English listeners. Our results for Igbo speakers listening to Igbo show that in general, the SNRs for Igbo sentences are lower than for English/English and Igbo/English. This means that the Igbo listeners could understand 50% of the Igbo sentences at SNRs less than those required for English sentences by either native or non-native listeners. This result can be explained by the fact that the perception of Igbo utterances by Igbo subjects may have been aided by the prediction of tonal and vowel harmony features existent in the Igbo language. In agreement with other studies, our results also show that in a noisy environment listeners are able to perceive their native language better than a second language. The ability of native language speakers to perceive their language better than a second language in a noisy environment may be attributed to the fact that: Native speakers are more familiar with the sounds of their language than second language speakers.One of the features of language is that it is predictable hence even in noise a native speaker may be able to predict a succeeding word that is scarcely audible. These contextual effects are facilitated by familiarity.

  16. 77 FR 60138 - Trinity Adaptive Management Working Group; Public Teleconference/Web-Based Meeting

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-02

    ... statements must be supplied to Elizabeth Hadley in one of the following formats: One hard copy with original... file formats are Adobe Acrobat PDF, MS Word, PowerPoint, or rich text file). Registered speakers who...

  17. 75 FR 29998 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-05-28

    ... Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification to fulfill the.... English, DSCA/DBO/CFM, (703) 601-3740. The following is a copy of a letter to the Speaker of the House of...

  18. Intentional switching in auditory selective attention: Exploring age-related effects in a spatial setup requiring speech perception.

    PubMed

    Oberem, Josefa; Koch, Iring; Fels, Janina

    2017-06-01

    Using a binaural-listening paradigm, age-related differences in the ability to intentionally switch auditory selective attention between two speakers, defined by their spatial location, were examined. Therefore 40 normal-hearing participants (20 young, Ø 24.8years; 20 older Ø 67.8years) were tested. The spatial reproduction of stimuli was provided by headphones using head-related-transfer-functions of an artificial head. Spoken number words of two speakers were presented simultaneously to participants from two out of eight locations on the horizontal plane. Guided by a visual cue indicating the spatial location of the target speaker, the participants were asked to categorize the target's number word into smaller vs. greater than five while ignoring the distractor's speech. Results showed significantly higher reaction times and error rates for older participants. The relative influence of the spatial switch of the target-speaker (switch or repetition of speaker's direction in space) was identical across age groups. Congruency effects (stimuli spoken by target and distractor may evoke the same answer or different answers) were increased for older participants and depend on the target's position. Results suggest that the ability to intentionally switch auditory attention to a new cued location was unimpaired whereas it was generally harder for older participants to suppress processing the distractor's speech. Copyright © 2017 Elsevier B.V. All rights reserved.

  19. Developing Appreciation for Sarcasm and Sarcastic Gossip: It Depends on Perspective

    ERIC Educational Resources Information Center

    Glenwright, Melanie; Tapley, Brent; Rano, Jacqueline K. S.; Pexman, Penny M.

    2017-01-01

    Background: Speakers use sarcasm to criticize others and to be funny; the indirectness of sarcasm protects the addressee's face (Brown & Levinson, 1987). Thus, appreciation of sarcasm depends on the ability to consider perspectives. Purpose: We investigated development of this ability from late childhood into adulthood and examined effects of…

  20. Absolute Interrogative Intonation Patterns in Buenos Aires Spanish

    ERIC Educational Resources Information Center

    Lee, Su Ar

    2010-01-01

    In Spanish, each uttered phrase, depending on its use, has one of a variety of intonation patterns. For example, a phrase such as "Maria viene manana" "Mary is coming tomorrow" can be used as a declarative or as an absolute interrogative (a yes/no question) depending on the intonation pattern that a speaker produces. …

  1. Investigating Executive Working Memory and Phonological Short-Term Memory in Relation to Fluency and Self-Repair Behavior in L2 Speech.

    PubMed

    Georgiadou, Effrosyni; Roehr-Brackin, Karen

    2017-08-01

    This paper reports the findings of a study investigating the relationship of executive working memory (WM) and phonological short-term memory (PSTM) to fluency and self-repair behavior during an unrehearsed oral task performed by second language (L2) speakers of English at two levels of proficiency, elementary and lower intermediate. Correlational analyses revealed a negative relationship between executive WM and number of pauses in the lower intermediate L2 speakers. However, no reliable association was found in our sample between executive WM or PSTM and self-repair behavior in terms of either frequency or type of self-repair. Taken together, our findings suggest that while executive WM may enhance performance at the conceptualization and formulation stages of the speech production process, self-repair behavior in L2 speakers may depend on factors other than working memory.

  2. Yemeni Arabic II.

    ERIC Educational Resources Information Center

    Qafisheh, Hamdi A.

    Instructional materials for advanced Sanaani Arabic, a dialect used predominantly for oral communication, include 25 units consisting of text derived from recordings of spontaneous conversations of native speakers in various communication situations. Some of the topics are: medical services, marriage, jobs, an interview, a car accident, proverbs…

  3. On the Development of Speech Resources for the Mixtec Language

    PubMed Central

    2013-01-01

    The Mixtec language is one of the main native languages in Mexico. In general, due to urbanization, discrimination, and limited attempts to promote the culture, the native languages are disappearing. Most of the information available about the Mixtec language is in written form as in dictionaries which, although including examples about how to pronounce the Mixtec words, are not as reliable as listening to the correct pronunciation from a native speaker. Formal acoustic resources, as speech corpora, are almost non-existent for the Mixtec, and no speech technologies are known to have been developed for it. This paper presents the development of the following resources for the Mixtec language: (1) a speech database of traditional narratives of the Mixtec culture spoken by a native speaker (labelled at the phonetic and orthographic levels by means of spectral analysis) and (2) a native speaker-adaptive automatic speech recognition (ASR) system (trained with the speech database) integrated with a Mixtec-to-Spanish/Spanish-to-Mixtec text translator. The speech database, although small and limited to a single variant, was reliable enough to build the multiuser speech application which presented a mean recognition/translation performance up to 94.36% in experiments with non-native speakers (the target users). PMID:23710134

  4. Speed-difficulty trade-off in speech: Chinese versus English

    PubMed Central

    Sun, Yao; Latash, Elizaveta M.; Mikaelian, Irina L.

    2011-01-01

    This study continues the investigation of the previously described speed-difficulty trade-off in picture description tasks. In particular, we tested a hypothesis that the Mandarin Chinese and American English are similar in showing logarithmic dependences between speech time and index of difficulty (ID), while they differ significantly in the amount of time needed to describe simple pictures, this difference increases for more complex pictures, and it is associated with a proportional difference in the number of syllables used. Subjects (eight Chinese speakers and eight English speakers) were tested in pairs. One subject (the Speaker) described simple pictures, while the other subject (the Performer) tried to reproduce the pictures based on the verbal description as quickly as possible with a set of objects. The Chinese speakers initiated speech production significantly faster than the English speakers. Speech time scaled linearly with ln(ID) in all subjects, but the regression coefficient was significantly higher in the English speakers as compared with the Chinese speakers. The number of errors was somewhat lower in the Chinese participants (not significantly). The Chinese pairs also showed a shorter delay between the initiation of speech and initiation of action by the Performer, shorter movement time by the Performer, and shorter overall performance time. The number of syllables scaled with ID, and the Chinese speakers used significantly smaller numbers of syllables. Speech rate was comparable between the two groups, about 3 syllables/s; it dropped for more complex pictures (higher ID). When asked to reproduce the same pictures without speaking, movement time scaled linearly with ln(ID); the Chinese performers were slower than the English performers. We conclude that natural languages show a speed-difficulty trade-off similar to Fitts’ law; the trade-offs in movement and speech production are likely to originate at a cognitive level. The time advantage of the Chinese participants originates not from similarity of the simple pictures and Chinese written characters and not from more sloppy performance. It is linked to using fewer syllables to transmit the same information. We suggest that natural languages may differ by informational density defined as the amount of information transmitted by a given number of syllables. PMID:21479658

  5. Words and pictures: An electrophysiological investigation of domain specific processing in native Chinese and English speakers

    PubMed Central

    Yum, Yen Na; Holcomb, Phillip J.; Grainger, Jonathan

    2011-01-01

    Comparisons of word and picture processing using Event-Related Potentials (ERPs) are contaminated by gross physical differences between the two types of stimuli. In the present study, we tackle this problem by comparing picture processing with word processing in an alphabetic and a logographic script, that are also characterized by gross physical differences. Native Mandarin Chinese speakers viewed pictures (line drawings) and Chinese characters (Experiment 1), native English speakers viewed pictures and English words (Experiment 2), and naïve Chinese readers (native English speakers) viewed pictures and Chinese characters (Experiment 3) in a semantic categorization task. The varying pattern of differences in the ERPs elicited by pictures and words across the three experiments provided evidence for i) script-specific processing arising between 150–200 ms post-stimulus onset, ii) domain-specific but script-independent processing arising between 200–300 ms post-stimulus onset, and iii) processing that depended on stimulus meaningfulness in the N400 time window. The results are interpreted in terms of differences in the way visual features are mapped onto higher-level representations for pictures and words in alphabetic and logographic writing systems. PMID:21439991

  6. Sentence Recall by Children With SLI Across Two Nonmainstream Dialects of English

    PubMed Central

    McDonald, Janet L.; Seidel, Christy M.; Hegarty, Michael

    2016-01-01

    Purpose The inability to accurately recall sentences has proven to be a clinical marker of specific language impairment (SLI); this task yields moderate-to-high levels of sensitivity and specificity. However, it is not yet known if these results hold for speakers of dialects whose nonmainstream grammatical productions overlap with those that are produced at high rates by children with SLI. Method Using matched groups of 70 African American English speakers and 36 Southern White English speakers and dialect-strategic scoring, we examined children's sentence recall abilities as a function of their dialect and clinical status (SLI vs. typically developing [TD]). Results For both dialects, the SLI group earned lower sentence recall scores than the TD group with sensitivity and specificity values ranging from .80 to .94, depending on the analysis. Children with SLI, as compared with TD controls, manifested lower levels of verbatim recall, more ungrammatical recalls when the recall was not exact, and higher levels of error on targeted functional categories, especially those marking tense. Conclusion When matched groups are examined and dialect-strategic scoring is used, sentence recall yields moderate-to-high levels of diagnostic accuracy to identify SLI within speakers of nonmainstream dialects of English. PMID:26501934

  7. Combining MEDLINE and publisher data to create parallel corpora for the automatic translation of biomedical text

    PubMed Central

    2013-01-01

    Background Most of the institutional and research information in the biomedical domain is available in the form of English text. Even in countries where English is an official language, such as the United States, language can be a barrier for accessing biomedical information for non-native speakers. Recent progress in machine translation suggests that this technique could help make English texts accessible to speakers of other languages. However, the lack of adequate specialized corpora needed to train statistical models currently limits the quality of automatic translations in the biomedical domain. Results We show how a large-sized parallel corpus can automatically be obtained for the biomedical domain, using the MEDLINE database. The corpus generated in this work comprises article titles obtained from MEDLINE and abstract text automatically retrieved from journal websites, which substantially extends the corpora used in previous work. After assessing the quality of the corpus for two language pairs (English/French and English/Spanish) we use the Moses package to train a statistical machine translation model that outperforms previous models for automatic translation of biomedical text. Conclusions We have built translation data sets in the biomedical domain that can easily be extended to other languages available in MEDLINE. These sets can successfully be applied to train statistical machine translation models. While further progress should be made by incorporating out-of-domain corpora and domain-specific lexicons, we believe that this work improves the automatic translation of biomedical texts. PMID:23631733

  8. The Speaker Respoken: Material Rhetoric as Feminist Methodology.

    ERIC Educational Resources Information Center

    Collins, Vicki Tolar

    1999-01-01

    Presents a methodology based on the concept of "material rhetoric" that can help scholars avoid problems as they reclaim women's historical texts. Defines material rhetoric and positions it theoretically in relation to other methodologies, including bibliographical studies, reception theory, and established feminist methodologies. Illustrates…

  9. Speech to Text Translation for Malay Language

    NASA Astrophysics Data System (ADS)

    Al-khulaidi, Rami Ali; Akmeliawati, Rini

    2017-11-01

    The speech recognition system is a front end and a back-end process that receives an audio signal uttered by a speaker and converts it into a text transcription. The speech system can be used in several fields including: therapeutic technology, education, social robotics and computer entertainments. In most cases in control tasks, which is the purpose of proposing our system, wherein the speed of performance and response concern as the system should integrate with other controlling platforms such as in voiced controlled robots. Therefore, the need for flexible platforms, that can be easily edited to jibe with functionality of the surroundings, came to the scene; unlike other software programs that require recording audios and multiple training for every entry such as MATLAB and Phoenix. In this paper, a speech recognition system for Malay language is implemented using Microsoft Visual Studio C#. 90 (ninety) Malay phrases were tested by 10 (ten) speakers from both genders in different contexts. The result shows that the overall accuracy (calculated from Confusion Matrix) is satisfactory as it is 92.69%.

  10. Birth order and mortality in two ethno-linguistic groups: Register-based evidence from Finland.

    PubMed

    Saarela, Jan; Cederström, Agneta; Rostila, Mikael

    2016-06-01

    Previous research has documented an association between birth order and suicide, although no study has examined whether it depends on the cultural context. Our aim was to study the association between birth order and cause-specific mortality in Finland, and whether it varies by ethno-linguistic affiliation. We used data from the Finnish population register, representing a 5% random sample of all Finnish speakers and a 20% random sample of Swedish speakers, who lived in Finland in any year 1987-2011. For each person, there was a link to all children who were alive in 1987. In total, there were 254,059 siblings in 96,387 sibling groups, and 9797 deaths. We used Cox regressions stratified by each siblings group and estimated all-cause and cause-specific mortality risks during the period 1987-2011. In line with previous research from Sweden, deaths from suicide were significantly associated with birth order. As compared to first-born, second-born had a suicide risk of 1.27, third-born of 1.35, and fourth- or higher-born of 1.72, while other causes of death did not display an evident and consistent birth-order pattern. Results for the Finnish-speaking siblings groups were almost identical to those based on both ethno-linguistic groups. In the Swedish-speaking siblings groups, there was no increase in the suicide risk by birth order, but a statistically not significant tendency towards an association with other external causes of death and deaths from cardiovascular diseases. Our findings provided evidence for an association between birth order and suicide among Finnish speakers in Finland, while no such association was found for Swedish speakers, suggesting that the birth order effect might depend on the cultural context. Copyright © 2016 Elsevier Ltd. All rights reserved.

  11. 'All the better for not seeing you': effects of communicative context on the speech of an individual with acquired communication difficulties.

    PubMed

    Bruce, Carolyn; Braidwood, Ursula; Newton, Caroline

    2013-01-01

    Evidence shows that speakers adjust their speech depending on the demands of the listener. However, it is unclear whether people with acquired communication disorders can and do make similar adaptations. This study investigated the impact of different conversational settings on the intelligibility of a speaker with acquired communication difficulties. Twenty-eight assessors listened to recordings of the speaker reading aloud 40 words and 32 sentences to a listener who was either face-to-face or unseen. The speaker's ability to convey information was measured by the accuracy of assessors' orthographic transcriptions of the words and sentences. Assessors' scores were significantly higher in the unseen condition for the single word task particularly if they had heard the face-to-face condition first. Scores for the sentence task were significantly higher in the second presentation regardless of the condition. The results from this study suggest that therapy conducted in situations where the client is not able to see their conversation partner may encourage them to perform at a higher level and increase the clarity of their speech. Readers will be able to describe: (1) the range of conversational adjustments made by speakers without communication difficulties; (2) differences between these tasks in offering contextual information to the listener; and (3) the potential for using challenging communicative situations to improve the performance of adults with communication disorders. Copyright © 2013 Elsevier Inc. All rights reserved.

  12. AMERICAN ENGLISH FOR INTERNATIONAL BUSINESSMEN.

    ERIC Educational Resources Information Center

    COSTINETT, SANDRA; ROSS, GORDON

    DESIGNED FOR NON-ENGLISH SPEAKERS ON AN INTERMEDIATE-ADVANCED LEVEL OF ENGLISH, THIS TEXT PROVIDES FOR PRACTICE IN CONVERSATIONAL BUSINESS ENGLISH. EACH OF THE TWELVE UNITS IS BASED ON A SPECIFIC TOPIC--PHONE CALLS, TRAVEL ARRANGEMENTS, QUARTERLY REPORTS AND SALES, ORGANIZATION CHARTS, ORDERING, PLANT LAYOUT, INVESTMENTS, EMPLOYEE BENEFITS,…

  13. History Forum Addresses Creation/Evolution Controversy.

    ERIC Educational Resources Information Center

    Schweinsberg, John

    1997-01-01

    A series of programs entitled Creationism and Evolution: The History of a Controversy was presented at the University of Alabama in Huntsville. The controversy was addressed from an historical and sociological, rather than a scientific perspective. Speakers addressed the evolution of scientific creationism, ancient texts versus sedimentary rocks…

  14. Interpreting Indian English Expository Prose.

    ERIC Educational Resources Information Center

    Kachru, Yamuna

    1988-01-01

    A study was undertaken to demonstrate that expository prose written in Indian English exhibits certain characteristics determined by the sociocultural conventions of writing in the Indian tradition. These features of Indian English texts are often judged to be inappropriate by native speakers of North American and British English, and mistakenly…

  15. Ways of looking ahead: hierarchical planning in language production.

    PubMed

    Lee, Eun-Kyung; Brown-Schmidt, Sarah; Watson, Duane G

    2013-12-01

    It is generally assumed that language production proceeds incrementally, with chunks of linguistic structure planned ahead of speech. Extensive research has examined the scope of language production and suggests that the size of planned chunks varies across contexts (Ferreira & Swets, 2002; Wagner & Jescheniak, 2010). By contrast, relatively little is known about the structure of advance planning, specifically whether planning proceeds incrementally according to the surface structure of the utterance, or whether speakers plan according to the hierarchical relationships between utterance elements. In two experiments, we examine the structure and scope of lexical planning in language production using a picture description task. Analyses of speech onset times and word durations show that speakers engage in hierarchical planning such that structurally dependent lexical items are planned together and that hierarchical planning occurs for both direct and indirect dependencies. Copyright © 2013 Elsevier B.V. All rights reserved.

  16. Opposing and following responses in sensorimotor speech control: Why responses go both ways.

    PubMed

    Franken, Matthias K; Acheson, Daniel J; McQueen, James M; Hagoort, Peter; Eisner, Frank

    2018-06-04

    When talking, speakers continuously monitor and use the auditory feedback of their own voice to control and inform speech production processes. When speakers are provided with auditory feedback that is perturbed in real time, most of them compensate for this by opposing the feedback perturbation. But some responses follow the perturbation. In the present study, we investigated whether the state of the speech production system at perturbation onset may determine what type of response (opposing or following) is made. The results suggest that whether a perturbation-related response is opposing or following depends on ongoing fluctuations of the production system: The system initially responds by doing the opposite of what it was doing. This effect and the nontrivial proportion of following responses suggest that current production models are inadequate: They need to account for why responses to unexpected sensory feedback depend on the production system's state at the time of perturbation.

  17. 76 FR 53525 - Final Environmental Impact Statement for the Proposed Keystone XL Project; Public Meetings

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-08-26

    ..., depending on the number of people who sign up to speak. Speakers will be asked to state their name and any organization with which they are affiliated. Depending on attendance, it may not be possible for all those who... take into account a wide range of factors, including environmental, economic, energy security, foreign...

  18. Structural impact and crashworthiness. Volume 1

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Davies, G.A.O.

    1984-01-01

    This volume contains the keynote addresses of those speakers invited to the International Confernece on Structural Impact and Crashworthiness held at Imperial College, London, in 1984. The speakers represent authoritative views on topics covering the spectrum of impact and crashworthiness involving several materials. The theme of this book may be summarized as 'understanding/modelling/prediction.' Ultimately a crashworthy design depends on many conceptual decisions being correct in the initial design phase. The overall configuration of a structure may be paramount; the detail design of joints and so on has to enable the structure to exploit energy absorption; the fail-safe features must notmore » be prohibitively expensive.« less

  19. LiT"gloss"

    ERIC Educational Resources Information Center

    LeLoup, Jean; Ponterio, Robert

    2007-01-01

    Reading in a second language (L2) is important for a variety of reasons. Reading authentic materials, in particular, allows L2 learners to engage with native speaker content, and ultimately the target culture. One body of authentic materials is that of literary texts, including those of more traditional prose fiction, poetry, and drama, but also…

  20. Sex Education. Kaleidoscope.

    ERIC Educational Resources Information Center

    Boderick, Carlfred; And Others

    This publication attempts to keep a portion of the educational community in touch with current topics and trends. The present focus is on sex education, a topic of growing concern in educational circles. The texts of three addresses are presented, with audience questions and speaker responses. Dr. Carlfred Broderick gives an overview of sex…

  1. Speakeasy. English for Special Purposes Series: Autobody Repair. Vietnamese Translation Manual.

    ERIC Educational Resources Information Center

    Van-Tanh, Phan

    This guide to spoken usage with exercises is designed to enrich the English vocabulary of speakers of Vietnamese. Narrative portions of the text are in Vietnamese. Emphasis is on survival skills and automobile repair skills. Chapters include "Time,""Getting to Know You,""Eating Out,""Opening a Savings…

  2. Speakeasy. English for Special Purposes Series: Autobody Repair. Ilocano Translation Manual.

    ERIC Educational Resources Information Center

    Berzabal, Ofelia G.

    This guide to spoken usage with exercises is designed to enrich the English vocabulary of speakers of Ilocano. Narrative portions of the text are in Ilocano. Emphasis is on survival language skills and automobile repair skills. Chapters include "Time,""Getting to Know You,""Eating Out,""Opening a Savings…

  3. Speakeasy. English for Special Purposes Series: Autobody Repair. Korean Translation Manual.

    ERIC Educational Resources Information Center

    Park, Yong-Ok

    This guide to spoken usage with exercises is designed to enrich the English vocabulary of speakers of Korean. Narrative portions of the text are in Korean. Emphasis is on survival skills and automobile repair skills. Chapters include "Time,""Getting to Know You,""Eating Out,""Opening a Savings…

  4. Critical Text Analysis: Linking Language and Cultural Studies

    ERIC Educational Resources Information Center

    Wharton, Sue

    2011-01-01

    Many UK universities offer degree programmes in English Language specifically for non-native speakers of English. Such programmes typically include not only language development but also development in various areas of content knowledge. A challenge that arises is to design courses in different areas that mutually support each other, thus…

  5. 75 FR 47275 - 36(b)(1) Arms Sales Notifications

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-08-05

    ... Department of Defense is publishing the unclassified text of two section 36(b)(1) arms sales notifications to... of letters to the Speaker of the House of Representatives, Transmittals 10-32 and 10-35 with... of the House of Representatives, Transmittal 10-32 with attached transmittal, policy justification...

  6. 76 FR 40703 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-07-11

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated July 21, 1996... copy of a letter to the Speaker of the House of [[Page 40704

  7. From Spanish to Portuguese.

    ERIC Educational Resources Information Center

    Ulsh, Jack L.

    This text is designed for native speakers of American English who have a good command of Spanish and want to acquire a knowledge of Portuguese. The manual focuses on features of Portuguese pronunciation, grammar, and vocabulary that have similar but not always identical counterparts in Spanish. Correspondences which often create interference for…

  8. New Thinking in School Chemistry.

    ERIC Educational Resources Information Center

    Organisation for Economic Cooperation and Development, Paris (France).

    This report contains the text of addresses given at a seminar on the status and development of the teaching of secondary school chemistry held in 1960 by the Organization for European Economic Development (now the Organization for Economic Cooperation and Development). The speakers considered advances in theoretical chemistry since 1900, the…

  9. Adaptation to an electropalatograph palate: acoustic, impressionistic, and perceptual data.

    PubMed

    McLeod, Sharynne; Searl, Jeff

    2006-05-01

    The purpose of this study was to evaluate adaptation to the electropalatograph (EPG) from the perspective of consonant acoustics, listener perceptions, and speaker ratings. Seven adults with typical speech wore an EPG and pseudo-EPG palate over 2 days and produced syllables, read a passage, counted, and rated their adaptation to the palate. Consonant acoustics, listener ratings, and speaker ratings were analyzed. The spectral mean for the burst (/t/) and frication (/s/) was reduced for the first 60-120 min of wearing the pseudo-EPG palate. Temporal features (stop gap, frication, and syllable duration) were unaffected by wearing the pseudo-EPG palate. The EPG palate had a similar effect on consonant acoustics as the pseudo-EPG palate. Expert listener ratings indicated minimal to no change in speech naturalness or distortion from the pseudo-EPG or EPG palate. The sounds [see text] were most likely to be affected. Speaker self-ratings related to oral comfort, speech, tongue movement, appearance, and oral sensation were negatively affected by the presence of the palatal devices. Speakers detected a substantial difference when wearing a palatal device, but the effects on speech were minimal based on listener ratings. Spectral features of consonants were initially affected, although adaptation occurred. Wearing an EPG or pseudo-EPG palate for approximately 2 hr results in relatively normal-sounding speech with acoustic features similar to a no-palate condition.

  10. Influence of encoding focus and stereotypes on source monitoring event-related-potentials.

    PubMed

    Leynes, P Andrew; Nagovsky, Irina

    2016-01-01

    Source memory, memory for the origin of a memory, can be influenced by stereotypes and the information of focus during encoding processes. Participants studied words from two different speakers (male or female) using self-focus or other-focus encoding. Source judgments for the speaker׳s voice and Event-Related Potentials (ERPs) were recorded during test. Self-focus encoding increased dependence on stereotype information and the Late Posterior Negativity (LPN). The results link the LPN with an increase in systematic decision processes such as consulting prior knowledge to support an episodic memory judgment. In addition, other-focus encoding increased conditional source judgments and resulted in weaker old/new recognition relative to the self-focus encoding. The putative correlate of recollection (LPC) was absent during this condition and this was taken as evidence that recollection of partial information supported source judgments. Collectively, the results suggest that other-focus encoding changes source monitoring processing by altering the weight of specific memory features. Copyright © 2015 Elsevier B.V. All rights reserved.

  11. Acoustic and perceptual effects of overall F0 range in a lexical pitch accent distinction

    NASA Astrophysics Data System (ADS)

    Wade, Travis

    2002-05-01

    A speaker's overall fundamental frequency range is generally considered a variable, nonlinguistic element of intonation. This study examined the precision with which overall F0 is predictable based on previous intonational context and the extent to which it may be perceptually significant. Speakers of Tokyo Japanese produced pairs of sentences differing lexically only in the presence or absence of a single pitch accent as responses to visual and prerecorded speech cues presented in an interactive manner. F0 placement of high tones (previously observed to be relatively variable in pitch contours) was found to be consistent across speakers and uniformly dependent on the intonation of the different sentences used as cues. In a subsequent perception experiment, continuous manipulation of these same sentences between typical accented and typical non-accent-containing versions were presented to Japanese listeners for lexical identification. Results showed that listeners' perception was not significantly altered in compensation for artificial manipulation of preceding intonation. Implications are discussed within an autosegmental analysis of tone. The current results are consistent with the notion that pitch range (i.e., specific vertical locations of tonal peaks) does not simply vary gradiently across speakers and situations but constitutes a predictable part of the phonetic specification of tones.

  12. The Impact of Dysphonic Voices on Healthy Listeners: Listener Reaction Times, Speech Intelligibility, and Listener Comprehension.

    PubMed

    Evitts, Paul M; Starmer, Heather; Teets, Kristine; Montgomery, Christen; Calhoun, Lauren; Schulze, Allison; MacKenzie, Jenna; Adams, Lauren

    2016-11-01

    There is currently minimal information on the impact of dysphonia secondary to phonotrauma on listeners. Considering the high incidence of voice disorders with professional voice users, it is important to understand the impact of a dysphonic voice on their audiences. Ninety-one healthy listeners (39 men, 52 women; mean age = 23.62 years) were presented with speech stimuli from 5 healthy speakers and 5 speakers diagnosed with dysphonia secondary to phonotrauma. Dependent variables included processing speed (reaction time [RT] ratio), speech intelligibility, and listener comprehension. Voice quality ratings were also obtained for all speakers by 3 expert listeners. Statistical results showed significant differences between RT ratio and number of speech intelligibility errors between healthy and dysphonic voices. There was not a significant difference in listener comprehension errors. Multiple regression analyses showed that voice quality ratings from the Consensus Assessment Perceptual Evaluation of Voice (Kempster, Gerratt, Verdolini Abbott, Barkmeier-Kraemer, & Hillman, 2009) were able to predict RT ratio and speech intelligibility but not listener comprehension. Results of the study suggest that although listeners require more time to process and have more intelligibility errors when presented with speech stimuli from speakers with dysphonia secondary to phonotrauma, listener comprehension may not be affected.

  13. Prosody and informativity: A cross-linguistic investigation

    NASA Astrophysics Data System (ADS)

    Ouyang, Iris Chuoying

    This dissertation aims to extend our knowledge of prosody -- in particular, what kinds of information may be conveyed through prosody, which prosodic dimensions may be used to convey them, and how individual speakers differ from one another in how they use prosody. Four production studies were conducted to examine how various factors interact with one another in shaping the prosody of an utterance and how prosody fulfills its multi-functional role. Experiments 1 explores the interaction between two types of informativity, namely information structure and information-theoretic properties. The results show that the prosodic consequences of new-information focus are modulated by the focused word's frequency, whereas the prosodic consequences of corrective focus are modulated by the focused word's probability in the context. Furthermore, f0 ranges appear to be more informative than f0 shapes in reflecting informativity across speakers. Specifically, speakers seem to have individual 'preferences' regarding f0 shapes, the f0 ranges they use for an utterance, and the magnitude of differences in f0 ranges by which they mark information-structural distinctions. In contrast, there is more cross-speaker validity in the actual directions of differences in f0 ranges between information-structural types. Experiments 2 and 3 further show that the interaction found between corrective focus and contextual probability depends on the interlocutor's knowledge state. When the interlocutor has no access to the crucial information concerning utterances' contextual probability, speakers prosodically emphasize contextually improbable corrections, but not contextually probable corrections. Furthermore, speakers prosodically emphasize the corrections in response to contextually probable misstatements, but not the corrections in response to contextually improbable misstatements. In contrast, completely opposite patterns are found when words' contextual probability is shared knowledge between the speaker and the interlocutor: speakers prosodically emphasize contextually probable corrections and the corrections in response to contextually improbable misstatements. Experiment 4 demonstrates the multi-functionality of prosody by investigating its discourse-level functions in Mandarin Chinese, a tone language where a word's prosodic patterns is crucial to its meaning. The results show that, although prosody serves fundamental, lexical-level functions in Mandarin Chinese, it nevertheless provides cues to information structure as well. Similar to what has been found with English, corrective information is prosodically more prominent than non-corrective information, and new information is prosodically more prominent than given information. Taken together, these experiments demonstrate the complex relationship between prosody and the different types of information it encodes in a given language. To better understand prosody, it is important to integrate insights from different traditions of research and to investigate across languages. In addition, the findings of this research suggest that speakers' assumptions about what their interlocutors know -- as well as speakers' ability to update these expectations -- play a key role in shaping the prosody of utterances. I hypothesize that prosodic prominence may reflect the gap between what speakers had expected their interlocutors to say and what their interlocutors have actually said.

  14. A positron emission tomography study of the neural basis of informational and energetic masking effects in speech perception

    NASA Astrophysics Data System (ADS)

    Scott, Sophie K.; Rosen, Stuart; Wickham, Lindsay; Wise, Richard J. S.

    2004-02-01

    Positron emission tomography (PET) was used to investigate the neural basis of the comprehension of speech in unmodulated noise (``energetic'' masking, dominated by effects at the auditory periphery), and when presented with another speaker (``informational'' masking, dominated by more central effects). Each type of signal was presented at four different signal-to-noise ratios (SNRs) (+3, 0, -3, -6 dB for the speech-in-speech, +6, +3, 0, -3 dB for the speech-in-noise), with listeners instructed to listen for meaning to the target speaker. Consistent with behavioral studies, there was SNR-dependent activation associated with the comprehension of speech in noise, with no SNR-dependent activity for the comprehension of speech-in-speech (at low or negative SNRs). There was, in addition, activation in bilateral superior temporal gyri which was associated with the informational masking condition. The extent to which this activation of classical ``speech'' areas of the temporal lobes might delineate the neural basis of the informational masking is considered, as is the relationship of these findings to the interfering effects of unattended speech and sound on more explicit working memory tasks. This study is a novel demonstration of candidate neural systems involved in the perception of speech in noisy environments, and of the processing of multiple speakers in the dorso-lateral temporal lobes.

  15. Linguistic Stereotyping in Older Adults' Perceptions of Health Care Aides.

    PubMed

    Rubin, Donald; Coles, Valerie Berenice; Barnett, Joshua Trey

    2016-07-01

    The cultural and linguistic diversity of the U.S. health care provider workforce is expanding. Diversity among health care personnel such as paraprofessional health care assistants (HCAs)-many of whom are immigrants-means that intimate, high-stakes cross-cultural and cross-linguistic contact characterizes many health interactions. In particular, nonmainstream HCAs may face negative patient expectations because of patients' language stereotypes. In other contexts, reverse linguistic stereotyping has been shown to result in negative speaker evaluations and even reduced listening comprehension quite independently of the actual language performance of the speaker. The present study extends the language and attitude paradigm to older adults' perceptions of HCAs. Listeners heard the identical speaker of Standard American English as they watched interactions between an HCA and an older patient. Ethnolinguistic identities-either an Anglo native speaker of English or a Mexican nonnative speaker-were ascribed to HCAs by means of fabricated personnel files. Dependent variables included measures of perceived HCA language proficiency, personal characteristics, and professional competence, as well as listeners' comprehension of a health message delivered by the putative HCA. For most of these outcomes, moderate effect sizes were found such that the HCA with an ascribed Anglo identity-relative to the Mexican guise-was judged more proficient in English, socially superior, interpersonally more attractive, more dynamic, and a more satisfactory home health aide. No difference in listening comprehension emerged, but the Anglo guise tended to engender a more compliant listening mind set. Results of this study can inform both provider-directed and patient-directed efforts to improve health care services for members of all linguistic and cultural groups.

  16. Berber Dialects. Materials Status Report.

    ERIC Educational Resources Information Center

    Center for Applied Linguistics, Washington, DC. Language/Area Reference Center.

    The materials status report for the Berber languages, minority languages spoken in northern Africa, is one of a series intended to provide the nonspecialist with a picture of the availability and quality of texts for teaching various languages to English speakers. The report consists of: (1) a brief narrative description of the Berber language,…

  17. Pausing, Preceding and Following "That" in English

    ERIC Educational Resources Information Center

    Bada, Erdogan

    2006-01-01

    While reading or speaking, individuals break up sentences into "meaningful chunks." This is true of any individual with any language background. Failure to do so, in an L2 context, leads to idiosyncrasies, and may possibly create some comprehensibility problems. In this study, native and non-native speakers of English read an authentic text into a…

  18. Learners' Perceptions of Listening Comprehension Problems.

    ERIC Educational Resources Information Center

    Hasan, Ali S.

    2000-01-01

    Reports a study of listening problems encountered in the English-as-a-foreign-language classroom in the English for Specific Purposes Centre at Damascus University in Syria. Looks particularly at learner strategies, features of the listening text, characteristics of the speaker, attitudes of the listener, the task to be completed as a result of…

  19. Adolescent Second Language Literacy: Language-Culture, Literature, and Identity.

    ERIC Educational Resources Information Center

    McCafferty, Steven G.

    2002-01-01

    Suggests adolescent second language speakers of English can better understand the second culture as embedded in the language, and feel more comfortable with expressing their sense of self through language, by discussing language and its use in the contexts of literary texts. Examines elements of the interface between language, culture, and…

  20. The Measurement of the Complexity, Accuracy, and Fluency of Written Arabic

    ERIC Educational Resources Information Center

    Raish, Michael

    2017-01-01

    This study investigates the multicomponential nature of L2 Arabic writing by adapting a number of direct measures of linguistic "complexity," "accuracy," and "fluency" (CAF) to the measurement of written Arabic texts produced by L2 Arabic learners and Arabic Native Speakers (NSs). Previous studies of L2 writing have…

  1. Annual Pennsylvania Conference on Postsecondary Occupational Education: Programming Postsecondary Occupational Education. (Ninth, Pennsylvania State University, September 28-29, 1977).

    ERIC Educational Resources Information Center

    Martorana, S. V., Ed.; And Others

    This publication contains the text of the main presentations and the highlights of discussion groups from the Ninth Annual Pennsylvania Conference on Postsecondary Occupational Education. The conference theme was "Programming Postsecondary Occupational Education." Ewald Nyquist, the first speaker, delineated the problems faced by…

  2. Ewe Basic Course. Revised Version.

    ERIC Educational Resources Information Center

    Warburton, Irene; And Others

    The purpose of this beginning text in Ewe is to provide the student with an introduction to the structure of the Ewe language and "reasonable practice" in speaking. It is intended to be taught with the assistance of a native speaker of Ewe. Linguistic terminology is minimal. Suggested teaching time ranges from two semesters of class meetings of…

  3. Cognitive Clozing To Teach Them To Think.

    ERIC Educational Resources Information Center

    Viaggio, Sergio

    A cloze-type procedure can be used effectively to teach interpreters how to anticipate what the speaker will say, inferring communicative intention. The exercise uses a text from which words are deleted, not randomly as in the true cloze procedure, but in significant locations or contexts. The words or groups of words suppressed are progressively…

  4. Kurdish. Materials Status Report.

    ERIC Educational Resources Information Center

    Center for Applied Linguistics, Washington, DC. Language/Area Reference Center.

    The materials status report for Kurdish is one of a series intended to provide the nonspecialist with a picture of the availability and quality of texts for teaching a given language to English speakers. Each report consists of: (1) a brief narrative description of the language, the areas where it is spoken, its major dialects, its writing system,…

  5. What Did You Say? How Did You Say It? Linguistics Choices in Online Discussions

    ERIC Educational Resources Information Center

    Oskoz, Ana; Pérez-Broncano, Olimpia

    2016-01-01

    Following appraisal theory (Martin & White, 2005), which examines the linguistic resources through which texts/speakers express, negotiate, and naturalize particular intersubjective or ideological positions (White, 2015), this study addresses (1) the extent to which second language (L2) learners express their attitudes toward their own and the…

  6. 76 FR 30667 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-05-26

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... letter to the Speaker of the House of Representatives, Transmittals 11-04 with attached transmittal, and...

  7. 77 FR 60391 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-03

    ... Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR FURTHER... a copy of a letter to the Speaker of the House of Representatives, Transmittals 12-32 with attached...

  8. 77 FR 60384 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-03

    ... Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR FURTHER... a copy of a letter to the Speaker of the House of Representatives, Transmittals 12-54 with attached...

  9. 77 FR 60387 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-03

    ... Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR FURTHER... a copy of a letter to the Speaker of the House of Representatives, Transmittals 12-26 with attached...

  10. 76 FR 32958 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-06-07

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated July 21, 1996... of a letter to the Speaker of the House of Representatives, Transmittals 11-02 with attached...

  11. 76 FR 43659 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-07-21

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated July 21, 1996... of a letter to the Speaker of the House of Representatives, Transmittals 11-28 with attached...

  12. Research-Based Lessons That Support Student Independent Reading in Social Studies

    ERIC Educational Resources Information Center

    Swanson, Elizabeth; Reed, Deborah; Vaughn, Sharon

    2016-01-01

    High school social studies teachers face unique challenges in helping their students learn independently from text in their discipline. In this article, a set of research-based practices that couple independent student reading with high-quality instruction proven to improve content learning for high school nonnative English speakers is provided.…

  13. Studia Linguistica. A Journal of General Linguistics, Volume 35, Number 1-2.

    ERIC Educational Resources Information Center

    Sigurd, Bengt, Ed.; Svartvik, Jan, Ed.

    1981-01-01

    The proceedings of the 1981 International Congress of Applied Linguistics include the full text of papers read by invited speakers. The keynote address, four plenary lectures, seven special lectures, and one symposium symmary are presented. The papers cover such topics as the status of applied linguistics, bilingual education for majority and…

  14. Somali. Materials Status Report.

    ERIC Educational Resources Information Center

    Center for Applied Linguistics, Washington, DC. Language/Area Reference Center.

    The materials status report for Somali, the official language of Somalia, is one of a series intended to provide the nonspecialist with a picture of the availability and quality of texts for teaching various languages to English speakers. The report consists of: (1) a brief narrative description of Somali, the areas where it is spoken, its major…

  15. Supporting Oral Narrative Development of Kindergarten English Language Learners Using Multimedia Stories

    ERIC Educational Resources Information Center

    Yang, Sha

    2016-01-01

    Narrative ability comes before literacy for bilingual students and helps narrow down the gap in text-level literacy between English language learners (ELLs) and native English speakers. Kindergarten ELLs are the best age group to receive intervention to improve their oral narrative skills. Multimedia stories have potential to assist kindergarten…

  16. The Effect of Three Message Organization Variables Upon Listener Comprehension.

    ERIC Educational Resources Information Center

    Johnson, Arlee W.

    Public speaking texts urge speakers to organize their message in order to increase their audience's comprehension of it. Tests were run to determine if listeners understand better when three message organization variables are employed in a speech: explicit statement of the central idea, explicit statement of the main points, and transitions before…

  17. Family Structure & Social Change: A Preparation for Further Study Course.

    ERIC Educational Resources Information Center

    Donovan, Cathy

    This instructional unit, which is intended for Australians working toward a Certificate in General Education for Adults, contains activities to help learners develop the skills and knowledge to read and write complex texts while examining human relationships and the family. Aimed at both native and nonnative English speakers, the unit contains…

  18. The Digichaint Interactive Game as a Virtual Learning Environment for Irish

    ERIC Educational Resources Information Center

    Ní Chiaráin, Neasa; Ní Chasaide, Ailbhe

    2016-01-01

    Although Text-To-Speech (TTS) synthesis has been little used in Computer-Assisted Language Learning (CALL), it is ripe for deployment, particularly for minority and endangered languages, where learners have little access to native speaker models and where few genuinely interactive and engaging teaching/learning materials are available. These…

  19. Elaborer un exercice de grammaire (Working Out a Grammar Exercise)

    ERIC Educational Resources Information Center

    Principaud, Jeanne-Marie

    1977-01-01

    An elaboration of the official instruction on teaching French to native speakers in elementary school. The topics covered are: Methodological development of exercises; the linguistic ability and milieu of the students; operative criteria; and the question of a logical progression or spontaneous use of grammar exercises. (Text is in French.) (AMH)

  20. Translating the Verbal to the Visual

    ERIC Educational Resources Information Center

    Engbers, Susanna Kelly

    2012-01-01

    Communication has always been at least partly a visual experience--insofar as the speaker's appearance on a stage or the text's appearance on the page. Certainly, however, the experience is becoming more and more visual. Thus, equipping students with the tools necessary to analyze and evaluate the visual rhetoric that surrounds everyone is a task…

  1. Chinese-Mandarin: Basic Course. Volume XII: Lessons 120-131.

    ERIC Educational Resources Information Center

    Defense Language Inst., Monterey, CA.

    This is the twelfth of 16 volumes of audiolingual classroom instruction in Mandarin Chinese. The course is designed to train native English speakers to Level 3 Foreign Service Institute proficiency in comprehension and speaking, and to Level 2 proficiency in reading and writing Mandarin. Facility in reading, writing, and translating texts written…

  2. Speaker Recognition by Combining MFCC and Phase Information in Noisy Conditions

    NASA Astrophysics Data System (ADS)

    Wang, Longbiao; Minami, Kazue; Yamamoto, Kazumasa; Nakagawa, Seiichi

    In this paper, we investigate the effectiveness of phase for speaker recognition in noisy conditions and combine the phase information with mel-frequency cepstral coefficients (MFCCs). To date, almost speaker recognition methods are based on MFCCs even in noisy conditions. For MFCCs which dominantly capture vocal tract information, only the magnitude of the Fourier Transform of time-domain speech frames is used and phase information has been ignored. High complement of the phase information and MFCCs is expected because the phase information includes rich voice source information. Furthermore, some researches have reported that phase based feature was robust to noise. In our previous study, a phase information extraction method that normalizes the change variation in the phase depending on the clipping position of the input speech was proposed, and the performance of the combination of the phase information and MFCCs was remarkably better than that of MFCCs. In this paper, we evaluate the robustness of the proposed phase information for speaker identification in noisy conditions. Spectral subtraction, a method skipping frames with low energy/Signal-to-Noise (SN) and noisy speech training models are used to analyze the effect of the phase information and MFCCs in noisy conditions. The NTT database and the JNAS (Japanese Newspaper Article Sentences) database added with stationary/non-stationary noise were used to evaluate our proposed method. MFCCs outperformed the phase information for clean speech. On the other hand, the degradation of the phase information was significantly smaller than that of MFCCs for noisy speech. The individual result of the phase information was even better than that of MFCCs in many cases by clean speech training models. By deleting unreliable frames (frames having low energy/SN), the speaker identification performance was improved significantly. By integrating the phase information with MFCCs, the speaker identification error reduction rate was about 30%-60% compared with the standard MFCC-based method.

  3. Effect of the loss of auditory feedback on segmental parameters of vowels of postlingually deafened speakers.

    PubMed

    Schenk, Barbara S; Baumgartner, Wolf Dieter; Hamzavi, Jafar Sasan

    2003-12-01

    The most obvious and best documented changes in speech of postlingually deafened speakers are the rate, fundamental frequency, and volume (energy). These changes are due to the lack of auditory feedback. But auditory feedback affects not only the suprasegmental parameters of speech. The aim of this study was to determine the change at the segmental level of speech in terms of vowel formants. Twenty-three postlingually deafened and 18 normally hearing speakers were recorded reading a German text. The frequencies of the first and second formants and the vowel spaces of selected vowels in word-in-context condition were compared. All first formant frequencies (F1) of the postlingually deafened speakers were significantly different from those of the normally hearing people. The values of F1 were higher for the vowels /e/ (418+/-61 Hz compared with 359+/-52 Hz, P=0.006) and /o/ (459+/-58 compared with 390+/-45 Hz, P=0.0003) and lower for /a/ (765+/-115 Hz compared with 851+/-146 Hz, P=0.038). The second formant frequency (F2) only showed a significant increase for the vowel/e/(2016+/-347 Hz compared with 2279+/-250 Hz, P=0.012). The postlingually deafened people were divided into two subgroups according to duration of deafness (shorter/longer than 10 years of deafness). There was no significant difference in formant changes between the two groups. Our report demonstrated an effect of auditory feedback also on segmental features of speech of postlingually deafened people.

  4. Visual analysis of online social media to open up the investigation of stance phenomena

    PubMed Central

    Kucher, Kostiantyn; Schamp-Bjerede, Teri; Kerren, Andreas; Paradis, Carita; Sahlgren, Magnus

    2015-01-01

    Online social media are a perfect text source for stance analysis. Stance in human communication is concerned with speaker attitudes, beliefs, feelings and opinions. Expressions of stance are associated with the speakers' view of what they are talking about and what is up for discussion and negotiation in the intersubjective exchange. Taking stance is thus crucial for the social construction of meaning. Increased knowledge of stance can be useful for many application fields such as business intelligence, security analytics, or social media monitoring. In order to process large amounts of text data for stance analyses, linguists need interactive tools to explore the textual sources as well as the processed data based on computational linguistics techniques. Both original texts and derived data are important for refining the analyses iteratively. In this work, we present a visual analytics tool for online social media text data that can be used to open up the investigation of stance phenomena. Our approach complements traditional linguistic analysis techniques and is based on the analysis of utterances associated with two stance categories: sentiment and certainty. Our contributions include (1) the description of a novel web-based solution for analyzing the use and patterns of stance meanings and expressions in human communication over time; and (2) specialized techniques used for visualizing analysis provenance and corpus overview/navigation. We demonstrate our approach by means of text media on a highly controversial scandal with regard to expressions of anger and provide an expert review from linguists who have been using our tool. PMID:29249903

  5. Visual analysis of online social media to open up the investigation of stance phenomena.

    PubMed

    Kucher, Kostiantyn; Schamp-Bjerede, Teri; Kerren, Andreas; Paradis, Carita; Sahlgren, Magnus

    2016-04-01

    Online social media are a perfect text source for stance analysis. Stance in human communication is concerned with speaker attitudes, beliefs, feelings and opinions. Expressions of stance are associated with the speakers' view of what they are talking about and what is up for discussion and negotiation in the intersubjective exchange. Taking stance is thus crucial for the social construction of meaning. Increased knowledge of stance can be useful for many application fields such as business intelligence, security analytics, or social media monitoring. In order to process large amounts of text data for stance analyses, linguists need interactive tools to explore the textual sources as well as the processed data based on computational linguistics techniques. Both original texts and derived data are important for refining the analyses iteratively. In this work, we present a visual analytics tool for online social media text data that can be used to open up the investigation of stance phenomena. Our approach complements traditional linguistic analysis techniques and is based on the analysis of utterances associated with two stance categories: sentiment and certainty. Our contributions include (1) the description of a novel web-based solution for analyzing the use and patterns of stance meanings and expressions in human communication over time; and (2) specialized techniques used for visualizing analysis provenance and corpus overview/navigation. We demonstrate our approach by means of text media on a highly controversial scandal with regard to expressions of anger and provide an expert review from linguists who have been using our tool.

  6. Domain-specific impairment of source memory following a right posterior medial temporal lobe lesion.

    PubMed

    Peters, Jan; Koch, Benno; Schwarz, Michael; Daum, Irene

    2007-01-01

    This single case analysis of memory performance in a patient with an ischemic lesion affecting posterior but not anterior right medial temporal lobe (MTL) indicates that source memory can be disrupted in a domain-specific manner. The patient showed normal recognition memory for gray-scale photos of objects (visual condition) and spoken words (auditory condition). While memory for visual source (texture/color of the background against which pictures appeared) was within the normal range, auditory source memory (male/female speaker voice) was at chance level, a performance pattern significantly different from the control group. This dissociation is consistent with recent fMRI evidence of anterior/posterior MTL dissociations depending upon the nature of source information (visual texture/color vs. auditory speaker voice). The findings are in good agreement with the view of dissociable memory processing by the perirhinal cortex (anterior MTL) and parahippocampal cortex (posterior MTL), depending upon the neocortical input that these regions receive. (c) 2007 Wiley-Liss, Inc.

  7. Beyond the language given: the neural correlates of inferring speaker meaning.

    PubMed

    Bašnáková, Jana; Weber, Kirsten; Petersson, Karl Magnus; van Berkum, Jos; Hagoort, Peter

    2014-10-01

    Even though language allows us to say exactly what we mean, we often use language to say things indirectly, in a way that depends on the specific communicative context. For example, we can use an apparently straightforward sentence like "It is hard to give a good presentation" to convey deeper meanings, like "Your talk was a mess!" One of the big puzzles in language science is how listeners work out what speakers really mean, which is a skill absolutely central to communication. However, most neuroimaging studies of language comprehension have focused on the arguably much simpler, context-independent process of understanding direct utterances. To examine the neural systems involved in getting at contextually constrained indirect meaning, we used functional magnetic resonance imaging as people listened to indirect replies in spoken dialog. Relative to direct control utterances, indirect replies engaged dorsomedial prefrontal cortex, right temporo-parietal junction and insula, as well as bilateral inferior frontal gyrus and right medial temporal gyrus. This suggests that listeners take the speaker's perspective on both cognitive (theory of mind) and affective (empathy-like) levels. In line with classic pragmatic theories, our results also indicate that currently popular "simulationist" accounts of language comprehension fail to explain how listeners understand the speaker's intended message. © The Author 2013. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  8. Native Language Experience Shapes Neural Basis of Addressed and Assembled Phonologies

    PubMed Central

    Mei, Leilei; Xue, Gui; Lu, Zhong-Lin; He, Qinghua; Wei, Miao; Zhang, Mingxia; Dong, Qi; Chen, Chuansheng

    2015-01-01

    Previous studies have suggested differential engagement of addressed and assembled phonologies in reading Chinese and alphabetic languages (e.g., English) and the modulatory role of native language in learning to read a second language. However, it is not clear whether native language experience shapes the neural mechanisms of addressed and assembled phonologies. To address this question, we trained native Chinese and native English speakers to read the same artificial language (based on Korean Hangul) either through addressed (i.e., whole-word mapping) or assembled (i.e., grapheme-to-phoneme mapping) phonology. We found that, for both native Chinese and native English speakers, addressed phonology relied on the regions in the ventral pathway, whereas assembled phonology depended on the regions in the dorsal pathway. More importantly, we found that the neural mechanisms of addressed and assembled phonologies were shaped by native language experience. Specifically, two key regions for addressed phonology (i.e., the left middle temporal gyrus and right inferior temporal gyrus) showed greater activation for addressed phonology in native Chinese speakers, while one key region for assembled phonology (i.e., the left supramarginal gyrus) showed more activation for assembled phonology in native English speakers. These results provide direct neuroimaging evidence for the effect of native language experience on the neural mechanisms of phonological access in a new language and support the assimilation-accommodation hypothesis. PMID:25858447

  9. Vocal Age Disguise: The Role of Fundamental Frequency and Speech Rate and Its Perceived Effects.

    PubMed

    Skoog Waller, Sara; Eriksson, Mårten

    2016-01-01

    The relationship between vocal characteristics and perceived age is of interest in various contexts, as is the possibility to affect age perception through vocal manipulation. A few examples of such situations are when age is staged by actors, when ear witnesses make age assessments based on vocal cues only or when offenders (e.g., online groomers) disguise their voice to appear younger or older. This paper investigates how speakers spontaneously manipulate two age related vocal characteristics ( f 0 and speech rate) in attempt to sound younger versus older than their true age, and if the manipulations correspond to actual age related changes in f 0 and speech rate (Study 1). Further aims of the paper is to determine how successful vocal age disguise is by asking listeners to estimate the age of generated speech samples (Study 2) and to examine whether or not listeners use f 0 and speech rate as cues to perceived age. In Study 1, participants from three age groups (20-25, 40-45, and 60-65 years) agreed to read a short text under three voice conditions. There were 12 speakers in each age group (six women and six men). They used their natural voice in one condition, attempted to sound 20 years younger in another and 20 years older in a third condition. In Study 2, 60 participants (listeners) listened to speech samples from the three voice conditions in Study 1 and estimated the speakers' age. Each listener was exposed to all three voice conditions. The results from Study 1 indicated that the speakers increased fundamental frequency ( f 0 ) and speech rate when attempting to sound younger and decreased f 0 and speech rate when attempting to sound older. Study 2 showed that the voice manipulations had an effect in the sought-after direction, although the achieved mean effect was only 3 years, which is far less than the intended effect of 20 years. Moreover, listeners used speech rate, but not f 0 , as a cue to speaker age. It was concluded that age disguise by voice can be achieved by naïve speakers even though the perceived effect was smaller than intended.

  10. Speaker's voice as a memory cue.

    PubMed

    Campeanu, Sandra; Craik, Fergus I M; Alain, Claude

    2015-02-01

    Speaker's voice occupies a central role as the cornerstone of auditory social interaction. Here, we review the evidence suggesting that speaker's voice constitutes an integral context cue in auditory memory. Investigation into the nature of voice representation as a memory cue is essential to understanding auditory memory and the neural correlates which underlie it. Evidence from behavioral and electrophysiological studies suggest that while specific voice reinstatement (i.e., same speaker) often appears to facilitate word memory even without attention to voice at study, the presence of a partial benefit of similar voices between study and test is less clear. In terms of explicit memory experiments utilizing unfamiliar voices, encoding methods appear to play a pivotal role. Voice congruency effects have been found when voice is specifically attended at study (i.e., when relatively shallow, perceptual encoding takes place). These behavioral findings coincide with neural indices of memory performance such as the parietal old/new recollection effect and the late right frontal effect. The former distinguishes between correctly identified old words and correctly identified new words, and reflects voice congruency only when voice is attended at study. Characterization of the latter likely depends upon voice memory, rather than word memory. There is also evidence to suggest that voice effects can be found in implicit memory paradigms. However, the presence of voice effects appears to depend greatly on the task employed. Using a word identification task, perceptual similarity between study and test conditions is, like for explicit memory tests, crucial. In addition, the type of noise employed appears to have a differential effect. While voice effects have been observed when white noise is used at both study and test, using multi-talker babble does not confer the same results. In terms of neuroimaging research modulations, characterization of an implicit memory effect reflective of voice congruency is currently lacking. Copyright © 2014 Elsevier B.V. All rights reserved.

  11. Did you or I say pretty, rude or brief? An ERP study of the effects of speaker's identity on emotional word processing.

    PubMed

    Pinheiro, Ana P; Rezaii, Neguine; Nestor, Paul G; Rauber, Andréia; Spencer, Kevin M; Niznikiewicz, Margaret

    2016-02-01

    During speech comprehension, multiple cues need to be integrated at a millisecond speed, including semantic information, as well as voice identity and affect cues. A processing advantage has been demonstrated for self-related stimuli when compared with non-self stimuli, and for emotional relative to neutral stimuli. However, very few studies investigated self-other speech discrimination and, in particular, how emotional valence and voice identity interactively modulate speech processing. In the present study we probed how the processing of words' semantic valence is modulated by speaker's identity (self vs. non-self voice). Sixteen healthy subjects listened to 420 prerecorded adjectives differing in voice identity (self vs. non-self) and semantic valence (neutral, positive and negative), while electroencephalographic data were recorded. Participants were instructed to decide whether the speech they heard was their own (self-speech condition), someone else's (non-self speech), or if they were unsure. The ERP results demonstrated interactive effects of speaker's identity and emotional valence on both early (N1, P2) and late (Late Positive Potential - LPP) processing stages: compared with non-self speech, self-speech with neutral valence elicited more negative N1 amplitude, self-speech with positive valence elicited more positive P2 amplitude, and self-speech with both positive and negative valence elicited more positive LPP. ERP differences between self and non-self speech occurred in spite of similar accuracy in the recognition of both types of stimuli. Together, these findings suggest that emotion and speaker's identity interact during speech processing, in line with observations of partially dependent processing of speech and speaker information. Copyright © 2016. Published by Elsevier Inc.

  12. Flexible spatial perspective-taking: conversational partners weigh multiple cues in collaborative tasks.

    PubMed

    Galati, Alexia; Avraamides, Marios N

    2013-01-01

    Research on spatial perspective-taking often focuses on the cognitive processes of isolated individuals as they adopt or maintain imagined perspectives. Collaborative studies of spatial perspective-taking typically examine speakers' linguistic choices, while overlooking their underlying processes and representations. We review evidence from two collaborative experiments that examine the contribution of social and representational cues to spatial perspective choices in both language and the organization of spatial memory. Across experiments, speakers organized their memory representations according to the convergence of various cues. When layouts were randomly configured and did not afford intrinsic cues, speakers encoded their partner's viewpoint in memory, if available, but did not use it as an organizing direction. On the other hand, when the layout afforded an intrinsic structure, speakers organized their spatial memories according to the person-centered perspective reinforced by the layout's structure. Similarly, in descriptions, speakers considered multiple cues whether available a priori or at the interaction. They used partner-centered expressions more frequently (e.g., "to your right") when the partner's viewpoint was misaligned by a small offset or coincided with the layout's structure. Conversely, they used egocentric expressions more frequently when their own viewpoint coincided with the intrinsic structure or when the partner was misaligned by a computationally difficult, oblique offset. Based on these findings we advocate for a framework for flexible perspective-taking: people weigh multiple cues (including social ones) to make attributions about the relative difficulty of perspective-taking for each partner, and adapt behavior to minimize their collective effort. This framework is not specialized for spatial reasoning but instead emerges from the same principles and memory-depended processes that govern perspective-taking in non-spatial tasks.

  13. Flexible spatial perspective-taking: conversational partners weigh multiple cues in collaborative tasks

    PubMed Central

    Galati, Alexia; Avraamides, Marios N.

    2013-01-01

    Research on spatial perspective-taking often focuses on the cognitive processes of isolated individuals as they adopt or maintain imagined perspectives. Collaborative studies of spatial perspective-taking typically examine speakers' linguistic choices, while overlooking their underlying processes and representations. We review evidence from two collaborative experiments that examine the contribution of social and representational cues to spatial perspective choices in both language and the organization of spatial memory. Across experiments, speakers organized their memory representations according to the convergence of various cues. When layouts were randomly configured and did not afford intrinsic cues, speakers encoded their partner's viewpoint in memory, if available, but did not use it as an organizing direction. On the other hand, when the layout afforded an intrinsic structure, speakers organized their spatial memories according to the person-centered perspective reinforced by the layout's structure. Similarly, in descriptions, speakers considered multiple cues whether available a priori or at the interaction. They used partner-centered expressions more frequently (e.g., “to your right”) when the partner's viewpoint was misaligned by a small offset or coincided with the layout's structure. Conversely, they used egocentric expressions more frequently when their own viewpoint coincided with the intrinsic structure or when the partner was misaligned by a computationally difficult, oblique offset. Based on these findings we advocate for a framework for flexible perspective-taking: people weigh multiple cues (including social ones) to make attributions about the relative difficulty of perspective-taking for each partner, and adapt behavior to minimize their collective effort. This framework is not specialized for spatial reasoning but instead emerges from the same principles and memory-depended processes that govern perspective-taking in non-spatial tasks. PMID:24133432

  14. Identifying intrinsic constituents of focus through ``imitation via restoration.''

    NASA Astrophysics Data System (ADS)

    Xu, Yi; Xu, Ching X.; Sun, Xuejing

    2003-04-01

    In this study we test the hypothesis that although certain parts of an observed intonation may seem dispensable in perception tests, they nevertheless are consistently produced by speakers. We refer to all consistently produced parts of an intonation as its ``intrinsic constituents.'' To identify the intrinsic constituents, we developed an experimental paradigm called ``imitation via restoration.'' In this paradigm, the intonation under scrutiny is first recorded by a native speaker. Then words carrying a potential constituent of the intonation are replaced by a loud noise. During the experiment, the sentence containing the replacement noise is presented to the subjects together with the text. The subjects' task is to repeat the sentence in exactly the same way as they hear it. The consistency with which subjects restore the missing parts of the target intonation would therefore provide a reasonable indication as to which of them are truly intrinsic to the intonation. Our first such experiment was conducted on determining whether focus consists of only on-focus pitch range expansion or it also involves obligatory post-focus pitch range suppression. Eight native speakers of Beijing Mandarin participated as subjects. Preliminary results have provided supporting evidence for the dual-component hypothesis.

  15. On the Right To Use the Language of One's Choice in Slovakia.

    ERIC Educational Resources Information Center

    Kontra, Miklos

    1997-01-01

    The text of a November 1995 Slovak Republic law concerning language use in that country is translated and analyzed from the perspective of a recent Linguistic Society of America (LSA) statement on language rights stating that speakers be allowed to express themselves, publicly or privately, in the language of their choice. The law provides that…

  16. Temps, aspect et modalisation dans un acte de communication (Tense, Aspect and Mood in an Act of Communication)

    ERIC Educational Resources Information Center

    Bentolila, Alain

    1977-01-01

    Syntactic analysis identifies the "conductor wires" permitting the listener to conceptualize the speaker's image of reality from information furnished by the linguistic message. Syntactic study is thus situated in the economy of the communication process. Mood, aspect and tense are studied in this framework. (Text is in French.) (AMH)

  17. Internet Health and the 21st-Century Patient: A Rhetorical View

    ERIC Educational Resources Information Center

    Segal, Judy Z.

    2009-01-01

    Internet health--here, the public use of information Web sites to facilitate decision making on matters of health and illness--is a rhetorical practice, involving text and trajectories of influence. A fulsome account of it requires attention to all parts of the rhetorical triangle--the speaker, the subject matter, and the audience--yet most…

  18. DAGUR MONGOLIAN GRAMMAR, TEXTS, AND LEXICON. URALIC AND ALTAIC SERIES, VOLUME 4.

    ERIC Educational Resources Information Center

    MARTIN, SAMUEL E.

    THIS DESCRIPTION OF DAGUR, A MONGOLIAN LANGUAGE, IS BASED ON THE DIALECT OF A SPEAKER BORN IN INNER MONGOLIA IN NORTHWEST MANCHURIA. SECTION I OF THIS MANUAL DESCRIBES OTHER WORKS PUBLISHED IN MONGOLIAN LINGUISTICS WHICH HAVE USED THE SAME INFORMANT, AND PRESENTS THE AUTHOR'S APPROACH, IN TERMS OF A PHRASE-STRUCTURE GRAMMAR. SECTION II PRESENTS…

  19. The Development of Composition Skills via Directed Writing.

    ERIC Educational Resources Information Center

    Rahilly, Leonard J.

    To alleviate problems associated with free composition as a method of foreign language writing instruction, the directed writing method was adapted for use in a college French composition course. High-quality French texts, often of only a page or two and written by native speakers, are used as a basis for grammatical analysis and discussion and a…

  20. Writing through Two Languages: First Language Expertise in a Language Minority Classroom

    ERIC Educational Resources Information Center

    Kibler, Amanda

    2010-01-01

    Language minority students' writing is often measured solely in terms of its distance from native speaker norms, yet doing so may ignore the process through which these texts are realized and the role that the first language plays in their creation. This study analyzes oral interactions among adolescent second language writers during an extended…

  1. L'application de l'appareil Suvaglingua de correction phonetique a l'enseignement de l'espagnol aux francophones (The Use of the Suvaglingua Synthesizer for Phonetic Correction in Spanish Courses for French Speakers)

    ERIC Educational Resources Information Center

    Sarmiento, Jose; And Others

    1974-01-01

    Describes the use of the verbo-tonal method of phonetic correction and the Suvaglingua synthesizer in Spanish courses at the International School of Interpreters at Mons, France. (Text is in French.) (PMP)

  2. 75 FR 76415 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-12-08

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 10-72 with...

  3. 75 FR 76418 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-12-08

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 10-65 with...

  4. 75 FR 76412 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-12-08

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 10-69 with...

  5. 76 FR 27022 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-05-10

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 11-12 with...

  6. 75 FR 76408 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-12-08

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 10-73 with...

  7. 76 FR 27026 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-05-10

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 11-13 with...

  8. 76 FR 17111 - 36(b)(1) Arms Sales Notification

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-03-28

    ... Department of Defense is publishing the unclassified text of a section 36(b)(1) arms sales notification. This is published to fulfill the requirements of section 155 of Public Law 104-164 dated 21 July 1996. FOR... following is a copy of a letter to the Speaker of the House of Representatives, Transmittals 11-10 with...

  9. Profiling the Collocation Use in ELT Textbooks and Learner Writing

    ERIC Educational Resources Information Center

    Tsai, Kuei-Ju

    2015-01-01

    The present study investigates the collocational profiles of (1) three series of graded textbooks for English as a foreign language (EFL) commonly used in Taiwan, (2) the written productions of EFL learners, and (3) the written productions of native speakers (NS) of English. These texts were examined against a purpose-built collocation list. Based…

  10. Comparing L1 and L2 Texts and Writers in First-Year Composition

    ERIC Educational Resources Information Center

    Eckstein, Grant; Ferris, Dana

    2018-01-01

    Scholars have at various points discussed the needs of second language (L2) writers enrolled in "mainstream" composition courses where they are mixed with native (L1) English speakers. Other researchers have investigated the experiences of L2 writers in mainstream classes and the perceptions of their instructors about their abilities and…

  11. Acquisition of L[subscript 2] Vocabulary for Effective Reading: Testing Teachers' Classroom Practice

    ERIC Educational Resources Information Center

    Llinares, Genoveva; Leiva, Bertha; Cartaya, Noela; St. Louis, Rubena

    2008-01-01

    Four practical approaches taken by teachers in their attempts to help students acquire the minimum sight vocabulary necessary for reading English for Science and Technology (EST) texts are described. Over a period of 12 weeks, subjects (native Spanish speakers at university level) were divided into 4 intact groups and worked under different…

  12. A Textbook of Modern Western Armenian.

    ERIC Educational Resources Information Center

    Bardakjian, Kevork B.; Thomson, Robert W.

    This textbook is designed as a first-year course at the undergraduate level. It is specifically aimed at students with no prior knowledge of Armenian, and emphasis has been placed on the numerous subtleties and nuances of Armenian which might seem strange or difficult to English speakers. In the text, stress has been put on explaining the grammar…

  13. SPOKEN AYACUCHO QUECHUA. UNITS 1-10.

    ERIC Educational Resources Information Center

    PARKER, GARY J.; SOLA, DONALD F.

    THIS BEGINNING COURSE IN AYACUCHO QUECHUA, SPOKEN BY ABOUT A MILLION PEOPLE IN SOUTH-CENTRAL PERU, WAS PREPARED TO INTRODUCE THE PHONOLOGY AND GRAMMAR OF THIS DIALECT TO SPEAKERS OF ENGLISH. THE FIRST OF TWO VOLUMES, IT SERVES AS A TEXT FOR A 6-WEEK INTENSIVE COURSE OF 20 CLASS HOURS A WEEK. THE AUTHORS COMPARE AND CONTRAST SIGNIFICANT FEATURES OF…

  14. Lecciones de Quechua. Primer Nivel, Segundo Nivel (Quechua Lessons. First Level, Second Level).

    ERIC Educational Resources Information Center

    Herrero, Joaquin; And Others

    The two volumes comprising this text represent the beginning and intermediate levels of a course in the Quechua language for speakers of Spanish. Based on the "Audio Lingual Materials" format, the ten lessons in the the first volume generally contain: (1) a short dialog and a supplementary vocabulary section or adapted dialog which…

  15. TU-D-213AB-01: How You Can Be the Speaker and Communicator Everyone Wants You to Be.

    PubMed

    Collins, J; Aydogan, B

    2012-06-01

    Effectiveness of an oral presentation depends on the ability of the speaker to communicate with the audience. An important part of this communication is focusing on two to five key points and emphasizing those points during the presentation. Every aspect of the presentation should be purposeful and directed at facilitating learners' achievement of the objectives. This necessitates that the speaker has carefully developed the objectives and built the presentation around attainment of the objectives. A presentation should be designed to include as much audience participation as possible, no matter the size of the audience. Techniques to encourage audience participation include questioning, brainstorming, small-group activities, role-playing, case-based examples, directed listening, and use of an audience response system. It is first necessary to motivate and gain attention of the learner for learning to take place. This can be accomplished through appropriate use of humor, anecdotes, and quotations. This course will review adult learning principles and effective presentation skills, Learning Objectives: 1. Apply adult learning principles. 2. Demonstrate effective presentations skills. © 2012 American Association of Physicists in Medicine.

  16. General perceptual contributions to lexical tone normalization.

    PubMed

    Huang, Jingyuan; Holt, Lori L

    2009-06-01

    Within tone languages that use pitch variations to contrast meaning, large variability exists in the pitches produced by different speakers. Context-dependent perception may help to resolve this perceptual challenge. However, whether speakers rely on context in contour tone perception is unclear; previous studies have produced inconsistent results. The present study aimed to provide an unambiguous test of the effect of context on contour lexical tone perception and to explore its underlying mechanisms. In three experiments, Mandarin listeners' perception of Mandarin first and second (high-level and mid-rising) tones was investigated with preceding speech and non-speech contexts. Results indicate that the mean fundamental frequency (f0) of a preceding sentence affects perception of contour lexical tones and the effect is contrastive. Following a sentence with a higher-frequency mean f0, the following syllable is more likely to be perceived as a lower frequency lexical tone and vice versa. Moreover, non-speech precursors modeling the mean spectrum of f0 also elicit this effect, suggesting general perceptual processing rather than articulatory-based or speaker-identity-driven mechanisms.

  17. Verification of endocrinological functions at a short distance between parametric speakers and the human body.

    PubMed

    Lee, Soomin; Katsuura, Tetsuo; Shimomura, Yoshihiro

    2011-01-01

    In recent years, a new type of speaker called the parametric speaker has been used to generate highly directional sound, and these speakers are now commercially available. In our previous study, we verified that the burden of the parametric speaker was lower than that of the general speaker for endocrine functions. However, nothing has yet been demonstrated about the effects of the shorter distance than 2.6 m between parametric speakers and the human body. Therefore, we investigated the distance effect on endocrinological function and subjective evaluation. Nine male subjects participated in this study. They completed three consecutive sessions: a 20-min quiet period as a baseline, a 30-min mental task period with general speakers or parametric speakers, and a 20-min recovery period. We measured salivary cortisol and chromogranin A (CgA) concentrations. Furthermore, subjects took the Kwansei-gakuin Sleepiness Scale (KSS) test before and after the task and also a sound quality evaluation test after it. Four experiments, one with a speaker condition (general speaker and parametric speaker), the other with a distance condition (0.3 m and 1.0 m), were conducted, respectively, at the same time of day on separate days. We used three-way repeated measures ANOVA (speaker factor × distance factor × time factor) to examine the effects of the parametric speaker. We found that the endocrinological functions were not significantly different between the speaker condition and the distance condition. The results also showed that the physiological burdens increased with progress in time independent of the speaker condition and distance condition.

  18. The Communication of Public Speaking Anxiety: Perceptions of Asian and American Speakers.

    ERIC Educational Resources Information Center

    Martini, Marianne; And Others

    1992-01-01

    Finds that U.S. audiences perceive Asian speakers to have more speech anxiety than U.S. speakers, even though Asian speakers do not self-report higher anxiety levels. Confirms that speech state anxiety is not communicated effectively between speakers and audiences for Asian or U.S. speakers. (SR)

  19. An Investigation of Syntactic Priming among German Speakers at Varying Proficiency Levels

    ERIC Educational Resources Information Center

    Ruf, Helena T.

    2011-01-01

    This dissertation investigates syntactic priming in second language (L2) development among three speaker populations: (1) less proficient L2 speakers; (2) advanced L2 speakers; and (3) LI speakers. Using confederate scripting this study examines how German speakers choose certain word orders in locative constructions (e.g., "Auf dem Tisch…

  20. Modeling Speaker Proficiency, Comprehensibility, and Perceived Competence in a Language Use Domain

    ERIC Educational Resources Information Center

    Schmidgall, Jonathan Edgar

    2013-01-01

    Research suggests that listener perceptions of a speaker's oral language use, or a speaker's "comprehensibility," may be influenced by a variety of speaker-, listener-, and context-related factors. Primary speaker factors include aspects of the speaker's proficiency in the target language such as pronunciation and…

  1. Do children go for the nice guys? The influence of speaker benevolence and certainty on selective word learning.

    PubMed

    Bergstra, Myrthe; DE Mulder, Hannah N M; Coopmans, Peter

    2018-04-06

    This study investigated how speaker certainty (a rational cue) and speaker benevolence (an emotional cue) influence children's willingness to learn words in a selective learning paradigm. In two experiments four- to six-year-olds learnt novel labels from two speakers and, after a week, their memory for these labels was reassessed. Results demonstrated that children retained the label-object pairings for at least a week. Furthermore, children preferred to learn from certain over uncertain speakers, but they had no significant preference for nice over nasty speakers. When the cues were combined, children followed certain speakers, even if they were nasty. However, children did prefer to learn from nice and certain speakers over nasty and certain speakers. These results suggest that rational cues regarding a speaker's linguistic competence trump emotional cues regarding a speaker's affective status in word learning. However, emotional cues were found to have a subtle influence on this process.

  2. Improvements of ModalMax High-Fidelity Piezoelectric Audio Device

    NASA Technical Reports Server (NTRS)

    Woodard, Stanley E.

    2005-01-01

    ModalMax audio speakers have been enhanced by innovative means of tailoring the vibration response of thin piezoelectric plates to produce a high-fidelity audio response. The ModalMax audio speakers are 1 mm in thickness. The device completely supplants the need to have a separate driver and speaker cone. ModalMax speakers can perform the same applications of cone speakers, but unlike cone speakers, ModalMax speakers can function in harsh environments such as high humidity or extreme wetness. New design features allow the speakers to be completely submersed in salt water, making them well suited for maritime applications. The sound produced from the ModalMax audio speakers has sound spatial resolution that is readily discernable for headset users.

  3. Context Effects on Lexical Choice and Lexical Activation

    ERIC Educational Resources Information Center

    Jescheniak, Jorg D.; Hantsch, Ansgar; Schriefers, Herbert

    2005-01-01

    Speakers are regularly confronted with the choice among lexical alternatives when referring to objects, including basic-level names (e.g., car) and subordinate-level names (e.g., Beetle). Which of these names is eventually selected often depends on contextual factors. The present article reports a series of picture-word interference experiments…

  4. Reconsidering Language Orientation for Undergraduate Singers

    ERIC Educational Resources Information Center

    Paver, Barbara E.

    2009-01-01

    Foreign language lyric diction is a compulsory subject in all undergraduate vocal performance degrees in universities. However, the effectiveness of its teaching depends on the capacity of students to absorb the material, for which many are largely unprepared, due to their lack of previous language study. Further, native speakers of North American…

  5. Effects of Acoustic Variability on Second Language Vocabulary Learning

    ERIC Educational Resources Information Center

    Barcroft, Joe; Sommers, Mitchell S.

    2005-01-01

    This study examined the effects of acoustic variability on second language vocabulary learning. English native speakers learned new words in Spanish. Exposure frequency to the words was constant. Dependent measures were accuracy and latency of picture-to-Spanish and Spanish-to-English recall. Experiment 1 compared presentation formats of neutral…

  6. Facets of Speaking Proficiency

    ERIC Educational Resources Information Center

    de Jong, Nivja H.; Steinel, Margarita P.; Florijn, Arjen F.; Schoonen, Rob; Hulstijn, Jan H.

    2012-01-01

    This study examined the componential structure of second-language (L2) speaking proficiency. Participants--181 L2 and 54 native speakers of Dutch--performed eight speaking tasks and six tasks tapping nine linguistic skills. Performance in the speaking tasks was rated on functional adequacy by a panel of judges and formed the dependent variable in…

  7. Teaching First Language Speakers to Communicate across Linguistic Difference: Addressing Attitudes, Comprehension, and Strategies

    ERIC Educational Resources Information Center

    Subtirelu, Nicholas Close; Lindemann, Stephanie

    2016-01-01

    While most research in applied linguistics has focused on second language (L2) speakers and their language capabilities, the success of interaction between such speakers and first language (L1) speakers also relies on the positive attitudes and communication skills of the L1 speakers. However, some research has suggested that many L1 speakers lack…

  8. Temporal and acoustic characteristics of Greek vowels produced by adults with cerebral palsy

    NASA Astrophysics Data System (ADS)

    Botinis, Antonis; Orfanidou, Ioanna; Fourakis, Marios; Fourakis, Marios

    2005-09-01

    The present investigation examined the temporal and spectral characteristics of Greek vowels as produced by speakers with intact (NO) versus cerebral palsy affected (CP) neuromuscular systems. Six NO and six CP native speakers of Greek produced the Greek vowels [i, e, a, o, u] in the first syllable of CVCV nonsense words in a short carrier phrase. Stress could be on either the first or second syllable. There were three female and three male speakers in each group. In terms of temporal characteristics, the results showed that: vowels produced by CP speakers were longer than vowels produced by NO speakers; stressed vowels were longer than unstressed vowels; vowels produced by female speakers were longer than vowels produced by male speakers. In terms of spectral characteristics the results showed that the vowel space of the CP speakers was smaller than that of the NO speakers. This is similar to the results recently reported by Liu et al. [J. Acoust. Soc. Am. 117, 3879-3889 (2005)] for CP speakers of Mandarin. There was also a reduction of the acoustic vowel space defined by unstressed vowels, but this reduction was much more pronounced in the vowel productions of CP speakers than NO speakers.

  9. Consistency between verbal and non-verbal affective cues: a clue to speaker credibility.

    PubMed

    Gillis, Randall L; Nilsen, Elizabeth S

    2017-06-01

    Listeners are exposed to inconsistencies in communication; for example, when speakers' words (i.e. verbal) are discrepant with their demonstrated emotions (i.e. non-verbal). Such inconsistencies introduce ambiguity, which may render a speaker to be a less credible source of information. Two experiments examined whether children make credibility discriminations based on the consistency of speakers' affect cues. In Experiment 1, school-age children (7- to 8-year-olds) preferred to solicit information from consistent speakers (e.g. those who provided a negative statement with negative affect), over novel speakers, to a greater extent than they preferred to solicit information from inconsistent speakers (e.g. those who provided a negative statement with positive affect) over novel speakers. Preschoolers (4- to 5-year-olds) did not demonstrate this preference. Experiment 2 showed that school-age children's ratings of speakers were influenced by speakers' affect consistency when the attribute being judged was related to information acquisition (speakers' believability, "weird" speech), but not general characteristics (speakers' friendliness, likeability). Together, findings suggest that school-age children are sensitive to, and use, the congruency of affect cues to determine whether individuals are credible sources of information.

  10. Inferring speaker attributes in adductor spasmodic dysphonia: ratings from unfamiliar listeners.

    PubMed

    Isetti, Derek; Xuereb, Linnea; Eadie, Tanya L

    2014-05-01

    To determine whether unfamiliar listeners' perceptions of speakers with adductor spasmodic dysphonia (ADSD) differ from control speakers on the parameters of relative age, confidence, tearfulness, and vocal effort and are related to speaker-rated vocal effort or voice-specific quality of life. Twenty speakers with ADSD (including 6 speakers with ADSD plus tremor) and 20 age- and sex-matched controls provided speech recordings, completed a voice-specific quality-of-life instrument (Voice Handicap Index; Jacobson et al., 1997), and rated their own vocal effort. Twenty listeners evaluated speech samples for relative age, confidence, tearfulness, and vocal effort using rating scales. Listeners judged speakers with ADSD as sounding significantly older, less confident, more tearful, and more effortful than control speakers (p < .01). Increased vocal effort was strongly associated with decreased speaker confidence (rs = .88-.89) and sounding more tearful (rs = .83-.85). Self-rated speaker effort was moderately related (rs = .45-.52) to listener impressions. Listeners' perceptions of confidence and tearfulness were also moderately associated with higher Voice Handicap Index scores (rs = .65-.70). Unfamiliar listeners judge speakers with ADSD more negatively than control speakers, with judgments extending beyond typical clinical measures. The results have implications for counseling and understanding the psychosocial effects of ADSD.

  11. Speaker Localisation Using Time Difference of Arrival

    DTIC Science & Technology

    2008-04-01

    School of Electrical and Electronic Engineering of the University of Adelaide. His area of expertise and interest is in Signal Processing including audio ...support of Theatre intelligence capabilities. His recent research interests include: information visualisation , text and data mining, and speech and...by: steering microphone arrays to improve the quality of audio pickup for recording, communication and transcription; enhancing the separation – and

  12. Cultura, Comunicacion e interaccion: Hacia el contexto total del lenguage y el hombre hispanicos (Culture, Communication and Interaction: Towards a Total Context of the Spanish Language and Speaker)

    ERIC Educational Resources Information Center

    Poyatos, Fernando

    1974-01-01

    Described the methodological problems in setting up a kinesic inventory. Concludes that it is highly unrealistic to study language by itself without analyzing the formal and semantic make-up of the triple basic structure of language-paralanguage-kinesics. (Text is in Spanish.) (DS)

  13. Applied Linguistics: A Survey for Language Teachers. Collier MacMillan Teacher's Library.

    ERIC Educational Resources Information Center

    Kehoe, Monika, Ed.

    This book, written for those who have no formal training in linguistics but who have an interest in language or language teaching, is meant as a practical text for use in introductory courses in Applied Linguistics. Its emphasis is on second language teaching, but it is also of interest to those who teach native speakers because it covers the…

  14. Developing and Testing EVALOE: A Tool for Assessing Spoken Language Teaching and Learning in the Classroom

    ERIC Educational Resources Information Center

    Gràcia, Marta; Vega, Fàtima; Galván-Bovaira, Maria José

    2015-01-01

    Broadly speaking, the teaching of spoken language in Spanish schools has not been approached in a systematic way. Changes in school practices are needed in order to allow all children to become competent speakers and to understand and construct oral texts that are appropriate in different contexts and for different audiences both inside and…

  15. Thematic Structure in Barack Obama's Press Conference: A Systemic Functional Grammar Study

    ERIC Educational Resources Information Center

    Kuswoyo, Heri

    2016-01-01

    This article looks into the theme--rheme pattern of presidential press conference that can be employed by speakers to organize the text in order to have a texture. Since a message should be conveyed in clause contextually and co-textually. Therefore, the objectives of this study are to analyze and describe the theme-rheme pattern employed in…

  16. Text Enhancement and the Acquisition of English Verbal Inflection "-s" by L1 Haitian Creole Speakers

    ERIC Educational Resources Information Center

    De Santis, Paulina

    2008-01-01

    This article contributes to the growing body of research investigating the effects of drawing learner attention to the problematic aspects of the linguistic input in the context of meaning-focused instruction. One specific approach to concentrate learner attention on form in the written input is known as textual enhancement. The pilot study…

  17. The Effect of Instruction Type and Dyadic or Individual Emulation on the Quality of Higher-Order Peer Feedback in EFL

    ERIC Educational Resources Information Center

    Van Steendam, Elke; Rijlaarsdam, Gert; Sercu, Lies; Van den Bergh, Huub

    2010-01-01

    Studies in peer feedback on written texts show that instruction in revision is necessary for the effectiveness of global feedback. Participants in the study were 247 university freshmen, native speakers of Dutch, who took the same Business English course, and were instructed a revision strategy following Schunk and Zimmerman's social cognitive…

  18. Information Technology for Training and Education (ITTE): Conference Keynote Addresses (Brisbane, Queensland, Australia, February 4-8, 1991).

    ERIC Educational Resources Information Center

    Queensland Univ., Brisbane (Australia).

    This volume presents the text of nine keynote addresses and an abstract of one other address presented at ITTE'91, an international forum and conference on Information Technology for Training and Education (ITTE). Each address is preceded by biographical information on the speaker and followed by a brief critique by a professional in information…

  19. Re-Presenting Subversive Songs: Applying Strategies for Invention and Arrangement to Nontraditional Speech Texts

    ERIC Educational Resources Information Center

    Charlesworth, Dacia

    2010-01-01

    Invention deals with the content of a speech, arrangement involves placing the content in an order that is most strategic, style focuses on selecting linguistic devices, such as metaphor, to make the message more appealing, memory assists the speaker in delivering the message correctly, and delivery ideally enables great reception of the message.…

  20. Reader Identity: A Case Study of Korean Graduate Students' Meaning Construction of an L2 Literary Text

    ERIC Educational Resources Information Center

    Shin, Changok; Riazantseva, Anastasia

    2015-01-01

    Grounded in constructivist theories of reading and informed by the contemporary theories of identity, this study explored how three Korean adult speakers of English as a foreign language (EFL) constructed meaning of the novel "The Catcher in the Rye", and how their identities mediated this process. Sources of data included think aloud…

  1. Focusing on Culture-Related Episodes in a Teletandem Interaction between a Brazilian and an American Student

    ERIC Educational Resources Information Center

    Zakir, Maisa A.; Funo, Ludmila B. A.; Telles, João A.

    2016-01-01

    Teletandem is a telecollaborative learning context that involves pairs of native (or competent) speakers of different languages interacting through voice, text and webcam image. Using Skype, each participant plays the role of learner for half an hour, speaking and practising the language of his/her partner. This paper focuses on a teletandem…

  2. Learner Interaction Management in an Avatar and Chat-Based Virtual World

    ERIC Educational Resources Information Center

    Peterson, Mark

    2006-01-01

    In this paper, I report on the findings of a study that investigated non-native speaker interaction in a three dimensional (3D) virtual world that incorporates avatars and text chat known as "Active Worlds." Analysis of the chat transcripts indicated that the 24 intermediate level EFL participants were able to undertake a variety of tasks through…

  3. A Protocol for Annotating Parser Differences. Research Report. ETS RR-16-02

    ERIC Educational Resources Information Center

    Bruno, James V.; Cahill, Aoife; Gyawali, Binod

    2016-01-01

    We present an annotation scheme for classifying differences in the outputs of syntactic constituency parsers when a gold standard is unavailable or undesired, as in the case of texts written by nonnative speakers of English. We discuss its automated implementation and the results of a case study that uses the scheme to choose a parser best suited…

  4. Sensory Intelligence for Extraction of an Abstract Auditory Rule: A Cross-Linguistic Study.

    PubMed

    Guo, Xiao-Tao; Wang, Xiao-Dong; Liang, Xiu-Yuan; Wang, Ming; Chen, Lin

    2018-02-21

    In a complex linguistic environment, while speech sounds can greatly vary, some shared features are often invariant. These invariant features constitute so-called abstract auditory rules. Our previous study has shown that with auditory sensory intelligence, the human brain can automatically extract the abstract auditory rules in the speech sound stream, presumably serving as the neural basis for speech comprehension. However, whether the sensory intelligence for extraction of abstract auditory rules in speech is inherent or experience-dependent remains unclear. To address this issue, we constructed a complex speech sound stream using auditory materials in Mandarin Chinese, in which syllables had a flat lexical tone but differed in other acoustic features to form an abstract auditory rule. This rule was occasionally and randomly violated by the syllables with the rising, dipping or falling tone. We found that both Chinese and foreign speakers detected the violations of the abstract auditory rule in the speech sound stream at a pre-attentive stage, as revealed by the whole-head recordings of mismatch negativity (MMN) in a passive paradigm. However, MMNs peaked earlier in Chinese speakers than in foreign speakers. Furthermore, Chinese speakers showed different MMN peak latencies for the three deviant types, which paralleled recognition points. These findings indicate that the sensory intelligence for extraction of abstract auditory rules in speech sounds is innate but shaped by language experience. Copyright © 2018 IBRO. Published by Elsevier Ltd. All rights reserved.

  5. Speaker Linking and Applications using Non-Parametric Hashing Methods

    DTIC Science & Technology

    2016-09-08

    clustering method based on hashing—canopy- clustering . We apply this method to a large corpus of speaker recordings, demonstrate performance tradeoffs...and compare to other hash- ing methods. Index Terms: speaker recognition, clustering , hashing, locality sensitive hashing. 1. Introduction We assume...speaker in our corpus. Second, given a QBE method, how can we perform speaker clustering —each clustering should be a single speaker, and a cluster should

  6. The effect of tonal changes on voice onset time in Mandarin esophageal speech.

    PubMed

    Liu, Hanjun; Ng, Manwa L; Wan, Mingxi; Wang, Supin; Zhang, Yi

    2008-03-01

    The present study investigated the effect of tonal changes on voice onset time (VOT) between normal laryngeal (NL) and superior esophageal (SE) speakers of Mandarin Chinese. VOT values were measured from the syllables /pha/, /tha/, and /kha/ produced at four tone levels by eight NL and seven SE speakers who were native speakers of Mandarin. Results indicated that Mandarin tones were associated with significantly different VOT values for NL speakers, in which high-falling tone was associated with significantly shorter VOT values than mid-rising tone and falling-rising tone. Regarding speaker group, SE speakers showed significantly shorter VOT values than NL speakers across all tone levels. This may be related to their use of pharyngoesophageal (PE) segment as another sound source. SE speakers appear to take a shorter time to start PE segment vibration compared to NL speakers using the vocal folds for vibration.

  7. Proficiency in English sentence stress production by Cantonese speakers who speak English as a second language (ESL).

    PubMed

    Ng, Manwa L; Chen, Yang

    2011-12-01

    The present study examined English sentence stress produced by native Cantonese speakers who were speaking English as a second language (ESL). Cantonese ESL speakers' proficiency in English stress production as perceived by English-speaking listeners was also studied. Acoustical parameters associated with sentence stress including fundamental frequency (F0), vowel duration, and intensity were measured from the English sentences produced by 40 Cantonese ESL speakers. Data were compared with those obtained from 40 native speakers of American English. The speech samples were also judged by eight native listeners who were native speakers of American English for placement, degree, and naturalness of stress. Results showed that Cantonese ESL speakers were able to use F0, vowel duration, and intensity to differentiate sentence stress patterns. Yet, both female and male Cantonese ESL speakers exhibited consistently higher F0 in stressed words than English speakers. Overall, Cantonese ESL speakers were found to be proficient in using duration and intensity to signal sentence stress, in a way comparable with English speakers. In addition, F0 and intensity were found to correlate closely with perceptual judgement and the degree of stress with the naturalness of stress.

  8. Salience Effects: L2 Sentence Production as a Window on L1 Speech Planning

    ERIC Educational Resources Information Center

    Antón-Méndez, Inés; Gerfen, Chip; Ramos, Miguel

    2016-01-01

    Salience influences grammatical structure during production in a language-dependent manner because different languages afford different options to satisfy preferences. During production, speakers may always try to satisfy all syntactic encoding preferences (e.g., salient entities to be mentioned early, themes to be assigned the syntactic function…

  9. Rapid Learning of Syllable Classes from a Perceptually Continuous Speech Stream

    ERIC Educational Resources Information Center

    Endress, Ansgar D.; Bonatti, Luca L.

    2007-01-01

    To learn a language, speakers must learn its words and rules from fluent speech; in particular, they must learn dependencies among linguistic classes. We show that when familiarized with a short artificial, subliminally bracketed stream, participants can learn relations about the structure of its words, which specify the classes of syllables…

  10. Speaker-dependent Multipitch Tracking Using Deep Neural Networks

    DTIC Science & Technology

    2015-01-01

    connections through time. Studies have shown that RNNs are good at modeling sequential data like handwriting [12] and speech [26]. We plan to explore RNNs in...Schmidhuber, and S. Fernández, “Unconstrained on-line handwriting recognition with recurrent neural networks,” in Proceedings of NIPS, 2008, pp. 577–584. [13

  11. A Usage-Based Approach to Preposition Placement in English as a Second Language

    ERIC Educational Resources Information Center

    Jach, Daniel

    2018-01-01

    This study examined the acquisition of preposition placement in English as a second language from a usage-based perspective. German and Chinese learners of English and English native speakers rated the acceptability of English oblique "wh" relative clauses in a magnitude estimation task. Results indicated that acceptability depended on…

  12. Code-Switching in Persian-English and Telugu-English Conversations: With a Focus on Light Verb Constructions

    ERIC Educational Resources Information Center

    Moradi, Hamzeh

    2014-01-01

    Depending on the demands of a particular communicative situation, bilingual or multilingual speakers ("bilingualism-multilingualism") will switch between language varieties. Code-switching is the practice of moving between variations of languages in different contexts. In an educational context, code-switching is defined as the practice…

  13. Applied Linguistics and the Use of Minority Languages in Education

    ERIC Educational Resources Information Center

    Cenoz, Jasone; Gorter, Durk

    2008-01-01

    Research on minority languages is ordinarily not well known by speakers of "big" languages but it has focused on several areas of Applied Linguistics and it is relevant to many areas. This current volume of "AILA Review" features five articles. Each of the articles emphasizes some aspects of research, depending on the recent…

  14. Encouraging Students to Engage with Native Speakers during Study Abroad

    ERIC Educational Resources Information Center

    Cadd, Marc

    2012-01-01

    Students, their parents, and educators trust that a study-abroad experience is the best way to increase linguistic proficiency. The professional literature, however, shows a much more complex picture. Gains in linguistic proficiency appear to depend on variables such as whether the students experience a homestay or dormitory, the length of time…

  15. Poorer Phonetic Perceivers Show Greater Benefit in Phonetic-Phonological Speech Learning

    ERIC Educational Resources Information Center

    Ingvalson, Erin M.; Barr, Allison M.; Wong, Patrick C. M.

    2013-01-01

    Purpose: Previous research has demonstrated that native English speakers can learn lexical tones in word context (pitch-to-word learning), to an extent. However, learning success depends on learners' pre-training sensitivity to pitch patterns. The aim of this study was to determine whether lexical pitch-pattern training given before lexical…

  16. Characteristics of Speaking Rate in the Dysarthria Associated with Amyotrophic Lateral Sclerosis.

    ERIC Educational Resources Information Center

    Turner, Greg S.; Weismer, Gary

    1993-01-01

    The ability to alter speaking rate was studied in nine adult subjects with amyotrophic lateral sclerosis and nine control subjects. Results suggest that the relationship between speaking rate, articulation rate, pause duration, and pause frequency remained largely intact for the dysarthric speakers. Data showed greater dependence on pausing by the…

  17. The Speaker Gender Gap at Critical Care Conferences.

    PubMed

    Mehta, Sangeeta; Rose, Louise; Cook, Deborah; Herridge, Margaret; Owais, Sawayra; Metaxa, Victoria

    2018-06-01

    To review women's participation as faculty at five critical care conferences over 7 years. Retrospective analysis of five scientific programs to identify the proportion of females and each speaker's profession based on conference conveners, program documents, or internet research. Three international (European Society of Intensive Care Medicine, International Symposium on Intensive Care and Emergency Medicine, Society of Critical Care Medicine) and two national (Critical Care Canada Forum, U.K. Intensive Care Society State of the Art Meeting) annual critical care conferences held between 2010 and 2016. Female faculty speakers. None. Male speakers outnumbered female speakers at all five conferences, in all 7 years. Overall, women represented 5-31% of speakers, and female physicians represented 5-26% of speakers. Nursing and allied health professional faculty represented 0-25% of speakers; in general, more than 50% of allied health professionals were women. Over the 7 years, Society of Critical Care Medicine had the highest representation of female (27% overall) and nursing/allied health professional (16-25%) speakers; notably, male physicians substantially outnumbered female physicians in all years (62-70% vs 10-19%, respectively). Women's representation on conference program committees ranged from 0% to 40%, with Society of Critical Care Medicine having the highest representation of women (26-40%). The female proportions of speakers, physician speakers, and program committee members increased significantly over time at the Society of Critical Care Medicine and U.K. Intensive Care Society State of the Art Meeting conferences (p < 0.05), but there was no temporal change at the other three conferences. There is a speaker gender gap at critical care conferences, with male faculty outnumbering female faculty. This gap is more marked among physician speakers than those speakers representing nursing and allied health professionals. Several organizational strategies can address this gender gap.

  18. Improving Speaker Recognition by Biometric Voice Deconstruction

    PubMed Central

    Mazaira-Fernandez, Luis Miguel; Álvarez-Marquina, Agustín; Gómez-Vilda, Pedro

    2015-01-01

    Person identification, especially in critical environments, has always been a subject of great interest. However, it has gained a new dimension in a world threatened by a new kind of terrorism that uses social networks (e.g., YouTube) to broadcast its message. In this new scenario, classical identification methods (such as fingerprints or face recognition) have been forcedly replaced by alternative biometric characteristics such as voice, as sometimes this is the only feature available. The present study benefits from the advances achieved during last years in understanding and modeling voice production. The paper hypothesizes that a gender-dependent characterization of speakers combined with the use of a set of features derived from the components, resulting from the deconstruction of the voice into its glottal source and vocal tract estimates, will enhance recognition rates when compared to classical approaches. A general description about the main hypothesis and the methodology followed to extract the gender-dependent extended biometric parameters is given. Experimental validation is carried out both on a highly controlled acoustic condition database, and on a mobile phone network recorded under non-controlled acoustic conditions. PMID:26442245

  19. Improving Speaker Recognition by Biometric Voice Deconstruction.

    PubMed

    Mazaira-Fernandez, Luis Miguel; Álvarez-Marquina, Agustín; Gómez-Vilda, Pedro

    2015-01-01

    Person identification, especially in critical environments, has always been a subject of great interest. However, it has gained a new dimension in a world threatened by a new kind of terrorism that uses social networks (e.g., YouTube) to broadcast its message. In this new scenario, classical identification methods (such as fingerprints or face recognition) have been forcedly replaced by alternative biometric characteristics such as voice, as sometimes this is the only feature available. The present study benefits from the advances achieved during last years in understanding and modeling voice production. The paper hypothesizes that a gender-dependent characterization of speakers combined with the use of a set of features derived from the components, resulting from the deconstruction of the voice into its glottal source and vocal tract estimates, will enhance recognition rates when compared to classical approaches. A general description about the main hypothesis and the methodology followed to extract the gender-dependent extended biometric parameters is given. Experimental validation is carried out both on a highly controlled acoustic condition database, and on a mobile phone network recorded under non-controlled acoustic conditions.

  20. Automatic speech recognition and training for severely dysarthric users of assistive technology: the STARDUST project.

    PubMed

    Parker, Mark; Cunningham, Stuart; Enderby, Pam; Hawley, Mark; Green, Phil

    2006-01-01

    The STARDUST project developed robust computer speech recognizers for use by eight people with severe dysarthria and concomitant physical disability to access assistive technologies. Independent computer speech recognizers trained with normal speech are of limited functional use by those with severe dysarthria due to limited and inconsistent proximity to "normal" articulatory patterns. Severe dysarthric output may also be characterized by a small mass of distinguishable phonetic tokens making the acoustic differentiation of target words difficult. Speaker dependent computer speech recognition using Hidden Markov Models was achieved by the identification of robust phonetic elements within the individual speaker output patterns. A new system of speech training using computer generated visual and auditory feedback reduced the inconsistent production of key phonetic tokens over time.

  1. Listeners' comprehension of uptalk in spontaneous speech.

    PubMed

    Tomlinson, John M; Fox Tree, Jean E

    2011-04-01

    Listeners' comprehension of phrase final rising pitch on declarative utterances, or uptalk, was examined to test the hypothesis that prolongations might differentiate conflicting functions of rising pitch. In Experiment 1 we found that listeners rated prolongations as indicating more speaker uncertainty, but that rising pitch was unrelated to ratings. In Experiment 2 we found that prolongations interacted with rising pitch when listeners monitored for words in the subsequent utterance. Words preceded by prolonged uptalk were monitored faster than words preceded by non-prolonged uptalk. In Experiment 3 we found that the interaction between rising pitch and prolongations depended on listeners' beliefs about speakers' mental states. Results support the theory that temporal and situational context are important in determining intonational meaning. Copyright © 2010 Elsevier B.V. All rights reserved.

  2. Reflecting on Native Speaker Privilege

    ERIC Educational Resources Information Center

    Berger, Kathleen

    2014-01-01

    The issues surrounding native speakers (NSs) and nonnative speakers (NNSs) as teachers (NESTs and NNESTs, respectively) in the field of teaching English to speakers of other languages (TESOL) are a current topic of interest. In many contexts, the native speaker of English is viewed as the model teacher, thus putting the NEST into a position of…

  3. English Speakers Attend More Strongly than Spanish Speakers to Manner of Motion when Classifying Novel Objects and Events

    ERIC Educational Resources Information Center

    Kersten, Alan W.; Meissner, Christian A.; Lechuga, Julia; Schwartz, Bennett L.; Albrechtsen, Justin S.; Iglesias, Adam

    2010-01-01

    Three experiments provide evidence that the conceptualization of moving objects and events is influenced by one's native language, consistent with linguistic relativity theory. Monolingual English speakers and bilingual Spanish/English speakers tested in an English-speaking context performed better than monolingual Spanish speakers and bilingual…

  4. Hybrid Speaker Recognition Using Universal Acoustic Model

    NASA Astrophysics Data System (ADS)

    Nishimura, Jun; Kuroda, Tadahiro

    We propose a novel speaker recognition approach using a speaker-independent universal acoustic model (UAM) for sensornet applications. In sensornet applications such as “Business Microscope”, interactions among knowledge workers in an organization can be visualized by sensing face-to-face communication using wearable sensor nodes. In conventional studies, speakers are detected by comparing energy of input speech signals among the nodes. However, there are often synchronization errors among the nodes which degrade the speaker recognition performance. By focusing on property of the speaker's acoustic channel, UAM can provide robustness against the synchronization error. The overall speaker recognition accuracy is improved by combining UAM with the energy-based approach. For 0.1s speech inputs and 4 subjects, speaker recognition accuracy of 94% is achieved at the synchronization error less than 100ms.

  5. Perceptual Detection of Subtle Dysphonic Traits in Individuals with Cervical Spinal Cord Injury Using an Audience Response Systems Approach.

    PubMed

    Johansson, Kerstin; Strömbergsson, Sofia; Robieux, Camille; McAllister, Anita

    2017-01-01

    Reduced respiratory function following lower cervical spinal cord injuries (CSCIs) may indirectly result in vocal dysfunction. Although self-reports indicate voice change and limitations following CSCI, earlier efforts using global perceptual ratings to distinguish speakers with CSCI from noninjured speakers have not been very successful. We investigate the use of an audience response system-based approach to distinguish speakers with CSCI from noninjured speakers, and explore whether specific vocal traits can be identified as characteristic for speakers with CSCI. Fourteen speech-language pathologists participated in a web-based perceptual task, where their overt reactions to vocal dysfunction were registered during the continuous playback of recordings of 36 speakers (18 with CSCI, and 18 matched controls). Dysphonic events were identified through manual perceptual analysis, to allow the exploration of connections between dysphonic events and listener reactions. More dysphonic events, and more listener reactions, were registered for speakers with CSCI than for noninjured speakers. Strain (particularly in phrase-final position) and creak (particularly in nonphrase-final position) distinguish speakers with CSCI from noninjured speakers. For the identification of intermittent and subtle signs of vocal dysfunction, an approach where the temporal distribution of symptoms is registered offers a viable means to distinguish speakers affected by voice dysfunction from non-affected speakers. In speakers with CSCI, clinicians should listen for presence of final strain and nonfinal creak, and pay attention to self-reported voice function and voice problems, to identify individuals in need for clinical assessment and intervention. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  6. Experiments on Urdu Text Recognition

    NASA Astrophysics Data System (ADS)

    Mukhtar, Omar; Setlur, Srirangaraj; Govindaraju, Venu

    Urdu is a language spoken in the Indian subcontinent by an estimated 130-270 million speakers. At the spoken level, Urdu and Hindi are considered dialects of a single language because of shared vocabulary and the similarity in grammar. At the written level, however, Urdu is much closer to Arabic because it is written in Nastaliq, the calligraphic style of the Persian-Arabic script. Therefore, a speaker of Hindi can understand spoken Urdu but may not be able to read written Urdu because Hindi is written in Devanagari script, whereas an Arabic writer can read the written words but may not understand the spoken Urdu. In this chapter we present an overview of written Urdu. Prior research in handwritten Urdu OCR is very limited. We present (perhaps) the first system for recognizing handwritten Urdu words. On a data set of about 1300 handwritten words, we achieved an accuracy of 70% for the top choice, and 82% for the top three choices.

  7. Logos Announced the Light of Salvation: Interpreting How John Presented His Message in John 1:1-18, According to Functional Grammar

    ERIC Educational Resources Information Center

    Pollinger, Seth

    2014-01-01

    This study of John 1:1-18 describes how John (the speaker) presented his message to his audience within their activity of verbal communication. By focusing on verbal meaning, this interpretation analyzes how John presented and expressed his meanings through language by interpreting this text based on the seamless interrelation between John's…

  8. Conference Highlights of the National Energy Education, Business and Labor Conference (Washington, D.C., January 15-17, 1979).

    ERIC Educational Resources Information Center

    1979

    Included in this document are texts of speeches given at the conference. The speakers whose remarks appear are: (1) John F. O'Leary, Deputy Secretary of the Department of Energy; (2) Cecil D. Andrus, Secretary of Interior; (3) Julian M. Carroll, Governor of Kentucky; (4) Arnold Packer, Assistant Secretary, Department of Labor; and (5) James…

  9. Defense Language Institute Russian Basic Course. Volumes XXVIII, Lessons 131-140. Volume XXX, Lessons 151-159.

    ERIC Educational Resources Information Center

    Defense Language Inst., Washington, DC.

    The 19 lessons in these two volumes are intended for the advanced phase of a 159-lesson intensive audiolingual basic Russian course developed recently by the Defense Language Institute to train native speakers of English to a Level 3 second language proficiency. These third and fifth volumes contain such features as (1) texts on the Russian Civil…

  10. Equity and Excellence. A Conference on Women's Studies and the Humanities (Greensboro, North Carolina, March 17-18, 1983). A Conference Report.

    ERIC Educational Resources Information Center

    Morgen, Sandra, Ed.; White, Judith, Ed.

    Proceedings of a conference to explore the impact of women's studies scholarship on the humanities are divided into eight sections. Section I, an introduction, describes the rationale of the conference and gives an overview of conference attendance. Section II introduces the keynote speakers. Section III contains the texts of the three keynote…

  11. Alternative Careers in Information/Library Services: Summary of Proceedings of a Workshop. Miscellaneous Studies No. 5.

    ERIC Educational Resources Information Center

    Minor, Barbara B., Ed.

    This report of a two-day workshop on non-traditional careers in librarianship provides the text of five major presentations as well as summaries of the three alternative sessions which followed. Each of the speakers described a different type of career: (1) Alice S. Warner of Warner-Eddison Associates is an information broker; (2) Robert E. Herz…

  12. Analysis of wolves and sheep. Final report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hogden, J.; Papcun, G.; Zlokarnik, I.

    1997-08-01

    In evaluating speaker verification systems, asymmetries have been observed in the ease with which people are able to break into other people`s voice locks. People who are good at breaking into voice locks are called wolves, and people whose locks are easy to break into are called sheep. (Goats are people that have a difficult time opening their own voice locks.) Analyses of speaker verification algorithms could be used to understand wolf/sheep asymmetries. Using the notion of a ``speaker space``, it is demonstrated that such asymmetries could arise even though the similarity of voice 1 to voice 2 is themore » same as the inverse similarity. This explains partially the wolf/sheep asymmetries, although there may be other factors. The speaker space can be computed from interspeaker similarity data using multidimensional scaling, and such speaker space can be used to given a good approximation of the interspeaker similarities. The derived speaker space can be used to predict which of the enrolled speakers are likely to be wolves and which are likely to be sheep. However, a speaker must first enroll in the speaker key system and then be compared to each of the other speakers; a good estimate of a person`s speaker space position could be obtained using only a speech sample.« less

  13. Investigating Auditory Processing of Syntactic Gaps with L2 Speakers Using Pupillometry

    ERIC Educational Resources Information Center

    Fernandez, Leigh; Höhle, Barbara; Brock, Jon; Nickels, Lyndsey

    2018-01-01

    According to the Shallow Structure Hypothesis (SSH), second language (L2) speakers, unlike native speakers, build shallow syntactic representations during sentence processing. In order to test the SSH, this study investigated the processing of a syntactic movement in both native speakers of English and proficient late L2 speakers of English using…

  14. A Model of Mandarin Tone Categories--A Study of Perception and Production

    ERIC Educational Resources Information Center

    Yang, Bei

    2010-01-01

    The current study lays the groundwork for a model of Mandarin tones based on both native speakers' and non-native speakers' perception and production. It demonstrates that there is variability in non-native speakers' tone productions and that there are differences in the perceptual boundaries in native speakers and non-native speakers. There…

  15. Literacy Skill Differences between Adult Native English and Native Spanish Speakers

    ERIC Educational Resources Information Center

    Herman, Julia; Cote, Nicole Gilbert; Reilly, Lenore; Binder, Katherine S.

    2013-01-01

    The goal of this study was to compare the literacy skills of adult native English and native Spanish ABE speakers. Participants were 169 native English speakers and 124 native Spanish speakers recruited from five prior research projects. The results showed that the native Spanish speakers were less skilled on morphology and passage comprehension…

  16. Grammatical Planning Units during Real-Time Sentence Production in Speakers with Agrammatic Aphasia and Healthy Speakers

    ERIC Educational Resources Information Center

    Lee, Jiyeon; Yoshida, Masaya; Thompson, Cynthia K.

    2015-01-01

    Purpose: Grammatical encoding (GE) is impaired in agrammatic aphasia; however, the nature of such deficits remains unclear. We examined grammatical planning units during real-time sentence production in speakers with agrammatic aphasia and control speakers, testing two competing models of GE. We queried whether speakers with agrammatic aphasia…

  17. Development of panel loudspeaker system: design, evaluation and enhancement.

    PubMed

    Bai, M R; Huang, T

    2001-06-01

    Panel speakers are investigated in terms of structural vibration and acoustic radiation. A panel speaker primarily consists of a panel and an inertia exciter. Contrary to conventional speakers, flexural resonance is encouraged such that the panel vibrates as randomly as possible. Simulation tools are developed to facilitate system integration of panel speakers. In particular, electro-mechanical analogy, finite element analysis, and fast Fourier transform are employed to predict panel vibration and the acoustic radiation. Design procedures are also summarized. In order to compare the panel speakers with the conventional speakers, experimental investigations were undertaken to evaluate frequency response, directional response, sensitivity, efficiency, and harmonic distortion of both speakers. The results revealed that the panel speakers suffered from a problem of sensitivity and efficiency. To alleviate the problem, a woofer using electronic compensation based on H2 model matching principle is utilized to supplement the bass response. As indicated in the result, significant improvement over the panel speaker alone was achieved by using the combined panel-woofer system.

  18. Word Durations in Non-Native English

    PubMed Central

    Baker, Rachel E.; Baese-Berk, Melissa; Bonnasse-Gahot, Laurent; Kim, Midam; Van Engen, Kristin J.; Bradlow, Ann R.

    2010-01-01

    In this study, we compare the effects of English lexical features on word duration for native and non-native English speakers and for non-native speakers with different L1s and a range of L2 experience. We also examine whether non-native word durations lead to judgments of a stronger foreign accent. We measured word durations in English paragraphs read by 12 American English (AE), 20 Korean, and 20 Chinese speakers. We also had AE listeners rate the `accentedness' of these non-native speakers. AE speech had shorter durations, greater within-speaker word duration variance, greater reduction of function words, and less between-speaker variance than non-native speech. However, both AE and non-native speakers showed sensitivity to lexical predictability by reducing second mentions and high frequency words. Non-native speakers with more native-like word durations, greater within-speaker word duration variance, and greater function word reduction were perceived as less accented. Overall, these findings identify word duration as an important and complex feature of foreign-accented English. PMID:21516172

  19. Experimental study on GMM-based speaker recognition

    NASA Astrophysics Data System (ADS)

    Ye, Wenxing; Wu, Dapeng; Nucci, Antonio

    2010-04-01

    Speaker recognition plays a very important role in the field of biometric security. In order to improve the recognition performance, many pattern recognition techniques have be explored in the literature. Among these techniques, the Gaussian Mixture Model (GMM) is proved to be an effective statistic model for speaker recognition and is used in most state-of-the-art speaker recognition systems. The GMM is used to represent the 'voice print' of a speaker through modeling the spectral characteristic of speech signals of the speaker. In this paper, we implement a speaker recognition system, which consists of preprocessing, Mel-Frequency Cepstrum Coefficients (MFCCs) based feature extraction, and GMM based classification. We test our system with TIDIGITS data set (325 speakers) and our own recordings of more than 200 speakers; our system achieves 100% correct recognition rate. Moreover, we also test our system under the scenario that training samples are from one language but test samples are from a different language; our system also achieves 100% correct recognition rate, which indicates that our system is language independent.

  20. Bystander capability to activate speaker function for continuous dispatcher assisted CPR in case of suspected cardiac arrest.

    PubMed

    Steensberg, Alvilda T; Eriksen, Mette M; Andersen, Lars B; Hendriksen, Ole M; Larsen, Heinrich D; Laier, Gunnar H; Thougaard, Thomas

    2017-06-01

    The European Resuscitation Council Guidelines 2015 recommend bystanders to activate their mobile phone speaker function, if possible, in case of suspected cardiac arrest. This is to facilitate continuous dialogue with the dispatcher including (if required) cardiopulmonary resuscitation instructions. The aim of this study was to measure the bystander capability to activate speaker function in case of suspected cardiac arrest. In 87days, a systematic prospective registration of bystander capability to activate the speaker function, when cardiac arrest was suspected, was performed. For those asked, "can you activate your mobile phone's speaker function", audio recordings were examined and categorized into groups according to the bystanders capability to activate speaker function on their own initiative, without instructions, or with instructions from the emergency medical dispatcher. Time delay was measured, in seconds, for the bystanders without pre-activated speaker function. 42.0% (58) was able to activate the speaker function without instructions, 2.9% (4) with instructions, 18.1% (25) on own initiative and 37.0% (51) were unable to activate the speaker function. The median time to activate speaker function was 19s and 8s, with and without instructions, respectively. Dispatcher assisted cardiopulmonary resuscitation with activated speaker function, in cases of suspected cardiac arrest, allows for continuous dialogue between the emergency medical dispatcher and the bystander. In this study, we found a 63.0% success rate of activating the speaker function in such situations. Copyright © 2017 Elsevier B.V. All rights reserved.

  1. Hemispheric lateralization of linguistic prosody recognition in comparison to speech and speaker recognition.

    PubMed

    Kreitewolf, Jens; Friederici, Angela D; von Kriegstein, Katharina

    2014-11-15

    Hemispheric specialization for linguistic prosody is a controversial issue. While it is commonly assumed that linguistic prosody and emotional prosody are preferentially processed in the right hemisphere, neuropsychological work directly comparing processes of linguistic prosody and emotional prosody suggests a predominant role of the left hemisphere for linguistic prosody processing. Here, we used two functional magnetic resonance imaging (fMRI) experiments to clarify the role of left and right hemispheres in the neural processing of linguistic prosody. In the first experiment, we sought to confirm previous findings showing that linguistic prosody processing compared to other speech-related processes predominantly involves the right hemisphere. Unlike previous studies, we controlled for stimulus influences by employing a prosody and speech task using the same speech material. The second experiment was designed to investigate whether a left-hemispheric involvement in linguistic prosody processing is specific to contrasts between linguistic prosody and emotional prosody or whether it also occurs when linguistic prosody is contrasted against other non-linguistic processes (i.e., speaker recognition). Prosody and speaker tasks were performed on the same stimulus material. In both experiments, linguistic prosody processing was associated with activity in temporal, frontal, parietal and cerebellar regions. Activation in temporo-frontal regions showed differential lateralization depending on whether the control task required recognition of speech or speaker: recognition of linguistic prosody predominantly involved right temporo-frontal areas when it was contrasted against speech recognition; when contrasted against speaker recognition, recognition of linguistic prosody predominantly involved left temporo-frontal areas. The results show that linguistic prosody processing involves functions of both hemispheres and suggest that recognition of linguistic prosody is based on an inter-hemispheric mechanism which exploits both a right-hemispheric sensitivity to pitch information and a left-hemispheric dominance in speech processing. Copyright © 2014 Elsevier Inc. All rights reserved.

  2. "I May Be a Native Speaker but I'm Not Monolingual": Reimagining "All" Teachers' Linguistic Identities in TESOL

    ERIC Educational Resources Information Center

    Ellis, Elizabeth M.

    2016-01-01

    Teacher linguistic identity has so far mainly been researched in terms of whether a teacher identifies (or is identified by others) as a native speaker (NEST) or nonnative speaker (NNEST) (Moussu & Llurda, 2008; Reis, 2011). Native speakers are presumed to be monolingual, and nonnative speakers, although by definition bilingual, tend to be…

  3. (abstract) Synthesis of Speaker Facial Movements to Match Selected Speech Sequences

    NASA Technical Reports Server (NTRS)

    Scott, Kenneth C.

    1994-01-01

    We are developing a system for synthesizing image sequences the simulate the facial motion of a speaker. To perform this synthesis, we are pursuing two major areas of effort. We are developing the necessary computer graphics technology to synthesize a realistic image sequence of a person speaking selected speech sequences. Next, we are developing a model that expresses the relation between spoken phonemes and face/mouth shape. A subject is video taped speaking an arbitrary text that contains expression of the full list of desired database phonemes. The subject is video taped from the front speaking normally, recording both audio and video detail simultaneously. Using the audio track, we identify the specific video frames on the tape relating to each spoken phoneme. From this range we digitize the video frame which represents the extreme of mouth motion/shape. Thus, we construct a database of images of face/mouth shape related to spoken phonemes. A selected audio speech sequence is recorded which is the basis for synthesizing a matching video sequence; the speaker need not be the same as used for constructing the database. The audio sequence is analyzed to determine the spoken phoneme sequence and the relative timing of the enunciation of those phonemes. Synthesizing an image sequence corresponding to the spoken phoneme sequence is accomplished using a graphics technique known as morphing. Image sequence keyframes necessary for this processing are based on the spoken phoneme sequence and timing. We have been successful in synthesizing the facial motion of a native English speaker for a small set of arbitrary speech segments. Our future work will focus on advancement of the face shape/phoneme model and independent control of facial features.

  4. Stromal-epithelial dynamics in response to fractionated radiotherapy

    NASA Astrophysics Data System (ADS)

    Rong, Panying

    The speech of individuals with velopharyngeal incompetency (VPI) is characterized by hypernasality, a speech quality related to excessive emission of acoustic energy through the nose, as caused by failure of velopharyngeal closure. As an attempt to reduce hypernasality and, in turn, improve the quality of VPI-related hypernasal speech, this study is dedicated to developing an approach that uses speech-dependent articulatory adjustments to reduce hypernasality caused by excessive velopharyngeal opening. A preliminary study has been done to derive such articulatory adjustments for hypernasal /i/ vowels based on the simulation of an articulatorymodel (Speech Processing and Synthesis Toolboxes, Childers (2000)). Both nasal /i/ vowels with and without articulatory adjustments were synthesized by the model. Spectral analysis found that nasal acoustic features were attenuated and oral formant structures were restored after articulatory adjustments. In addition, comparisons of perceptual ratings of nasality between the two types of nasal vowels showed the articulatory adjustments generated by the model significantly reduced the perception of nasality for nasal /i/ vowels. Such articulatory adjustments for nasal /i/ have two patterns: 1) a consistent adjustment pattern, which corresponds an expansion at the velopharynx, and 2) some speech-dependent fine-tuning adjustment patterns, including adjustments in the lip area and the upper pharynx. The long-term goal of this study is to apply this approach of articulatory adjustment as a therapeutic tool in clinical speech treatment to detect and correct the maladaptive articulatory behaviors developed spontaneously by speakers with VPI on individual bases. This study constructed a speaker-adaptive articulatory model on the basis of the framework of Childers's vocal tract model to simulate articulatory adjustments aiming at compensating for the acoustic outcome caused by velopharyngeal opening and reducing nasality. To construct such a speaker-adaptive articulatory model, (1) an articulatory-acoustic-aerodynamic database was recorded using the articulography and aerodynamic instruments to provide point-wise articulatory data to be fitted into the framework of Childers's standard vocal tract model; (2) the length and transverse dimension of the vocal tract were adjusted to fit individual speaker by minimizing the acoustic discrepancy between the model simulation and the target derived from acoustic signal in the database using the simulated annealing algorithm; (3) the articulatory space of the model was adjusted to fit individual articulatory features by adapting the movement ranges of all articulators. With the speaker-adaptive articulatory model, the articulatory configurations of the oral and nasal vowels in the database were simulated and synthesized. Given the acoustic targets derived from the oral vowels in the database, speech-dependent articulatory adjustments were simulated to compensate for the acoustic outcome caused by VPO. The resultant articulatory configurations corresponds to nasal vowels with articulatory adjustment, which were synthesized to serve as the perceptual stimuli for a listening task of nasality rating. The oral and nasal vowels synthesized based on the oral and nasal vowel targets in the database also served as the perceptual stimuli. The results suggest both acoustic and perceptual effects of the mode-generated articulatory adjustment on the nasal vowels /a/, /i/ and /u/. In terms of acoustics, the articulatory adjustment (1) restores the altered formant structures due to nasal coupling, including shifted formant frequency, attenuated formant intensity and expanded formant bandwidth and (2) attenuates the peaks and zeros caused by nasal resonances. Perceptually, the articulatory adjustment generated by the speaker-adaptive model significantly reduces the perceived nasality for all three vowels (/a/, /i/, /u/). The acoustic and perceptual effects of articulatory adjustment suggest achievement of the acoustic goal of compensating for the acoustic discrepancy caused by VPO and the auditory goal of reducing the perception of nasality. Such a finding is consistent with motor equivalence (Hughes and Abbs, 1976; Maeda, 1990), which enables inter-articulator coordination to compensate for the deviation from the acoustic/auditory goal caused by the shifted position of an articulator. The articulatory adjustment responsible for the acoustic and perceptual effects as described above was decomposed into a set of empirical orthogonal modes (Story and Titze, 1998). Both gross articulatory patterns and fine-tuning adjustments were found in the principal orthogonal modes, which lead to the acoustic compensation and reduction of nasality. For /a/ and /i/, a direct relationship was found among the acoustic features, nasality, and articulatory adjustment patterns. Specifically, the articulatory adjustments indicated by the principal orthogonal modes of the adjusted nasal /a/ and /i/ were directly correlated with the attenuation of the acoustic cues of nasality (i.e., shifting of F1 and F2 frequencies) and the reduction of nasality rating. For /u/, such a direct relationship among the acoustic features, nasality and articulatory adjustment was not as prominent, suggesting the possibility of additional acoustic correlates of nasality other than F1 and F2. The findings of this study demonstrate the possibility of using articulatory adjustment to reduce the perception of nasality through model simulation. A speaker-adaptive articulatory model is able to simulate individual-based articulatory adjustment strategies that can be applied in clinical settings to serve as the articulatory targets for correction of the maladaptive articulatory behaviors developed spontaneously by speakers with hypernasal speech. Such a speaker-adaptive articulatory model provides an intuitive way of articulatory learning and self-training for speakers with VPI to learn appropriate articulatory strategies through model-speaker interaction.

  5. How Cognitive Load Influences Speakers' Choice of Referring Expressions.

    PubMed

    Vogels, Jorrig; Krahmer, Emiel; Maes, Alfons

    2015-08-01

    We report on two experiments investigating the effect of an increased cognitive load for speakers on the choice of referring expressions. Speakers produced story continuations to addressees, in which they referred to characters that were either salient or non-salient in the discourse. In Experiment 1, referents that were salient for the speaker were non-salient for the addressee, and vice versa. In Experiment 2, all discourse information was shared between speaker and addressee. Cognitive load was manipulated by the presence or absence of a secondary task for the speaker. The results show that speakers under load are more likely to produce pronouns, at least when referring to less salient referents. We take this finding as evidence that speakers under load have more difficulties taking discourse salience into account, resulting in the use of expressions that are more economical for themselves. © 2014 Cognitive Science Society, Inc.

  6. Unsupervised real-time speaker identification for daily movies

    NASA Astrophysics Data System (ADS)

    Li, Ying; Kuo, C.-C. Jay

    2002-07-01

    The problem of identifying speakers for movie content analysis is addressed in this paper. While most previous work on speaker identification was carried out in a supervised mode using pure audio data, more robust results can be obtained in real-time by integrating knowledge from multiple media sources in an unsupervised mode. In this work, both audio and visual cues will be employed and subsequently combined in a probabilistic framework to identify speakers. Particularly, audio information is used to identify speakers with a maximum likelihood (ML)-based approach while visual information is adopted to distinguish speakers by detecting and recognizing their talking faces based on face detection/recognition and mouth tracking techniques. Moreover, to accommodate for speakers' acoustic variations along time, we update their models on the fly by adapting to their newly contributed speech data. Encouraging results have been achieved through extensive experiments, which shows a promising future of the proposed audiovisual-based unsupervised speaker identification system.

  7. The irreversibility of sensitive period effects in language development: evidence from second language acquisition in international adoptees.

    PubMed

    Norrman, Gunnar; Bylund, Emanuel

    2016-05-01

    The question of a sensitive period in language acquisition has been subject to extensive research and debate for more than half a century. While it has been well established that the ability to learn new languages declines in early years, the extent to which this outcome depends on biological maturation in contrast to previously acquired knowledge remains disputed. In the present study, we addressed this question by examining phonetic discriminatory abilities in early second language (L2) speakers of Swedish, who had either maintained their first language (L1) (immigrants) or had lost it (international adoptees), using native speaker controls. Through this design, we sought to disentangle the effects of the maturational state of the learner on L2 development from the effects of L1 interference: if additional language development is indeed constrained by an interfering L1, then adoptees should outperform immigrant speakers. The results of an auditory lexical decision task, in which fine vowel distinctions in Swedish had been modified, showed, however, no difference between the L2 groups. Instead, both L2 groups scored significantly lower than the native speaker group. The three groups did not differ in their ability to discriminate non-modified words. These findings demonstrate that L1 loss is not a crucial condition for successfully acquiring an L2, which in turn is taken as support for a maturational constraints view on L2 acquisition. A video abstract of this article can be viewed at: https://youtu.be/1J9X50aePeU. © 2015 John Wiley & Sons Ltd.

  8. Cross-language differences in the brain network subserving intelligible speech.

    PubMed

    Ge, Jianqiao; Peng, Gang; Lyu, Bingjiang; Wang, Yi; Zhuo, Yan; Niu, Zhendong; Tan, Li Hai; Leff, Alexander P; Gao, Jia-Hong

    2015-03-10

    How is language processed in the brain by native speakers of different languages? Is there one brain system for all languages or are different languages subserved by different brain systems? The first view emphasizes commonality, whereas the second emphasizes specificity. We investigated the cortical dynamics involved in processing two very diverse languages: a tonal language (Chinese) and a nontonal language (English). We used functional MRI and dynamic causal modeling analysis to compute and compare brain network models exhaustively with all possible connections among nodes of language regions in temporal and frontal cortex and found that the information flow from the posterior to anterior portions of the temporal cortex was commonly shared by Chinese and English speakers during speech comprehension, whereas the inferior frontal gyrus received neural signals from the left posterior portion of the temporal cortex in English speakers and from the bilateral anterior portion of the temporal cortex in Chinese speakers. Our results revealed that, although speech processing is largely carried out in the common left hemisphere classical language areas (Broca's and Wernicke's areas) and anterior temporal cortex, speech comprehension across different language groups depends on how these brain regions interact with each other. Moreover, the right anterior temporal cortex, which is crucial for tone processing, is equally important as its left homolog, the left anterior temporal cortex, in modulating the cortical dynamics in tone language comprehension. The current study pinpoints the importance of the bilateral anterior temporal cortex in language comprehension that is downplayed or even ignored by popular contemporary models of speech comprehension.

  9. Cross-language differences in the brain network subserving intelligible speech

    PubMed Central

    Ge, Jianqiao; Peng, Gang; Lyu, Bingjiang; Wang, Yi; Zhuo, Yan; Niu, Zhendong; Tan, Li Hai; Leff, Alexander P.; Gao, Jia-Hong

    2015-01-01

    How is language processed in the brain by native speakers of different languages? Is there one brain system for all languages or are different languages subserved by different brain systems? The first view emphasizes commonality, whereas the second emphasizes specificity. We investigated the cortical dynamics involved in processing two very diverse languages: a tonal language (Chinese) and a nontonal language (English). We used functional MRI and dynamic causal modeling analysis to compute and compare brain network models exhaustively with all possible connections among nodes of language regions in temporal and frontal cortex and found that the information flow from the posterior to anterior portions of the temporal cortex was commonly shared by Chinese and English speakers during speech comprehension, whereas the inferior frontal gyrus received neural signals from the left posterior portion of the temporal cortex in English speakers and from the bilateral anterior portion of the temporal cortex in Chinese speakers. Our results revealed that, although speech processing is largely carried out in the common left hemisphere classical language areas (Broca’s and Wernicke’s areas) and anterior temporal cortex, speech comprehension across different language groups depends on how these brain regions interact with each other. Moreover, the right anterior temporal cortex, which is crucial for tone processing, is equally important as its left homolog, the left anterior temporal cortex, in modulating the cortical dynamics in tone language comprehension. The current study pinpoints the importance of the bilateral anterior temporal cortex in language comprehension that is downplayed or even ignored by popular contemporary models of speech comprehension. PMID:25713366

  10. The Role of the Auditory Brainstem in Processing Linguistically-Relevant Pitch Patterns

    ERIC Educational Resources Information Center

    Krishnan, Ananthanarayan; Gandour, Jackson T.

    2009-01-01

    Historically, the brainstem has been neglected as a part of the brain involved in language processing. We review recent evidence of language-dependent effects in pitch processing based on comparisons of native vs. nonnative speakers of a tonal language from electrophysiological recordings in the auditory brainstem. We argue that there is enhancing…

  11. The Role of Mother Tongue Literacy in Language Learning and Mathematical Learning: Is There a Multilingual Benefit for Both?

    ERIC Educational Resources Information Center

    Dahm, Rebecca; De Angelis, Gessica

    2018-01-01

    The present study examines the multilingual benefit in relation to language learning and mathematical learning. The objective is to assess whether speakers of three or more languages, depending on language profile and personal histories, show significant advantages in language learning and/or mathematical learning, and whether mother tongue…

  12. The Effects of Learning English as a Second Language on the Acquisition of a New Phonemic Contrast.

    ERIC Educational Resources Information Center

    Streeter, Lynn A.; Landauer, Thomas K.

    Very sharp discrimination functions for the timing of voice onset relative to stop release characterize perceptual boundaries between certain pairs of stop consonants for adult speakers of many languages. To explore how these discriminations depend on experience, their development was studied among Kikuyu children, whose native language contains…

  13. Time to English Reading Proficiency. Research Brief. RB 1201

    ERIC Educational Resources Information Center

    Shneyderman, Aleksandr; Froman, Terry

    2012-01-01

    The time it takes for an English Language Learner (ELL) to reach reading proficiency in English depends on the grade level of entry into the English for Speakers of Other Languages (ESOL) program and on the student's initial English proficiency level. The summary table below presents the average years to English proficiency across different grade…

  14. Wednesday's Meeting Really Is on Friday: A Meta-Analysis and Evaluation of Ambiguous Spatiotemporal Language

    ERIC Educational Resources Information Center

    Stickles, Elise; Lewis, Tasha N.

    2018-01-01

    Experimental work has shown that spatial experiences influence spatiotemporal metaphor use. In these studies, participants are asked a question that yields different responses depending on the metaphor participants use. It has been claimed that English speakers are equally likely to respond with either variant in the absence of priming. Related…

  15. Argumentation and Participation in the Primary Mathematics Classroom: Two Episodes and Related Theoretical Abductions

    ERIC Educational Resources Information Center

    Krummheuer, Gotz

    2007-01-01

    The main assumption of this article is that learning mathematics depends on the student's participation in processes of collective argumentation. On the empirical level, such processes will be analyzed with Toulmin's theory of argumentation and Goffman's idea of decomposition of the speaker's role. On the theoretical level, different statuses of…

  16. Publishing Sami Literature--From Christian Translations to Sami Publishing Houses

    ERIC Educational Resources Information Center

    Paltto, Kirsti

    2010-01-01

    Publishing in the Sami languages has always been difficult. The Sami are currently spread across four countries, Norway, Sweden, Finland, and Russia. There are nine different Sami languages, some of them with only a few speakers. The Sami publishing industry is entirely dependent on government funding as it does not have its own funds nor is there…

  17. The Meaning of English Words across Cultures, with a Focus on Cameroon and Hong Kong

    ERIC Educational Resources Information Center

    Bobda, Augustin Simo

    2009-01-01

    A word, even when considered monosemic, generally has a cluster of meanings, depending on the mental representation of the referent by the speaker/writer or listener/reader. The variation is even more noticeable across cultures. This paper investigates the different ways in which cultural knowledge helps in the interpretation of English lexical…

  18. Local and State Relations: Proceedings of the Junior College Conference (Ocean Springs, Mississippi, June 24-26, 1968).

    ERIC Educational Resources Information Center

    Roberts, Dayton Y., Ed.

    Several speakers addressed this conference of the Southeastern Regional Junior College Leadership Program and the Mississippi Junior College Commission. (1) The balance of state and local relations (J.L.Wattenbarger) depends on current trends: the change from local to state or federal finance, national standards required by population mobility,…

  19. Contemplating Regretted Messages: Learning-Oriented, Repair-Oriented, and Emotion-Focused Reflection

    ERIC Educational Resources Information Center

    Meyer, Janet R.

    2013-01-01

    Regretted messages provide speakers an opportunity to learn. Whether learning occurs should depend upon how the incident is processed. This study had two objectives: (a) to determine how the goal a message conflicts with and seriousness influence the emotion(s) evoked; and (b) to determine which variables predict adoption of learning-oriented,…

  20. The Different Time Course of Phonotactic Constraint Learning in Children and Adults: Evidence from Speech Errors

    ERIC Educational Resources Information Center

    Smalle, Eleonore H. M.; Muylle, Merel; Szmalec, Arnaud; Duyck, Wouter

    2017-01-01

    Speech errors typically respect the speaker's implicit knowledge of language-wide phonotactics (e.g., /t/ cannot be a syllable onset in the English language). Previous work demonstrated that adults can learn novel experimentally induced phonotactic constraints by producing syllable strings in which the allowable position of a phoneme depends on…

  1. The Timing of Island Effects in Nonnative Sentence Processing

    ERIC Educational Resources Information Center

    Felser, Claudia; Cunnings, Ian; Batterham, Claire; Clahsen, Harald

    2012-01-01

    Using the eye-movement monitoring technique in two reading comprehension experiments, this study investigated the timing of constraints on wh-dependencies (so-called island constraints) in first- and second-language (L1 and L2) sentence processing. The results show that both L1 and L2 speakers of English are sensitive to extraction islands during…

  2. An EMA/EPG Study of Vowel-to-Vowel Articulation across Velars in Southern British English

    ERIC Educational Resources Information Center

    Fletcher, Janet

    2004-01-01

    Recent studies have attested that the extent of transconsonantal vowel-to-vowel coarticulation is at least partly dependent on degree of prosodic accentuation, in languages like English. A further important factor is the mutual compatibility of consonant and vowel gestures associated with the segments in question. In this study two speakers of…

  3. Long-Term Experience with Chinese Language Shapes the Fusiform Asymmetry of English Reading

    PubMed Central

    Mei, Leilei; Xue, Gui; Lu, Zhong-Lin; Chen, Chuansheng; Wei, Miao; He, Qinghua; Dong, Qi

    2015-01-01

    Previous studies have suggested differential engagement of the bilateral fusiform gyrus in the processing of Chinese and English. The present study tested the possibility that long-term experience with Chinese language affects the fusiform laterality of English reading by comparing three samples: Chinese speakers, English speakers with Chinese experience, and English speakers without Chinese experience. We found that, when reading words in their respective native language, Chinese and English speakers without Chinese experience differed in functional laterality of the posterior fusiform region (right laterality for Chinese speakers, but left laterality for English speakers). More importantly, compared with English speakers without Chinese experience, English speakers with Chinese experience showed more recruitment of the right posterior fusiform cortex for English words and pseudowords, which is similar to how Chinese speakers processed Chinese. These results suggest that long-term experience with Chinese shapes the fusiform laterality of English reading and have important implications for our understanding of the cross-language influences in terms of neural organization and of the functions of different fusiform subregions in reading. PMID:25598049

  4. The Effects of Self-Disclosure on Male and Female Perceptions of Individuals Who Stutter.

    PubMed

    Byrd, Courtney T; McGill, Megann; Gkalitsiou, Zoi; Cappellini, Colleen

    2017-02-01

    The purpose of this study was to examine the influence of self-disclosure on observers' perceptions of persons who stutter. Participants (N = 173) were randomly assigned to view 2 of 4 possible videos (i.e., male self-disclosure, male no self-disclosure, female self-disclosure, and female no self-disclosure). After viewing both videos, participants completed a survey assessing their perceptions of the speakers. Controlling for observer and speaker gender, listeners were more likely to select speakers who self-disclosed their stuttering as more friendly, outgoing, and confident compared with speakers who did not self-disclose. Observers were more likely to select speakers who did not self-disclose as unfriendly and shy compared with speakers who used a self-disclosure statement. Controlling for self-disclosure and observer gender, observers were less likely to choose the female speaker as friendlier, outgoing, and confident compared with the male speaker. Observers also were more likely to select the female speaker as unfriendly, shy, unintelligent, and insecure compared with the male speaker and were more likely to report that they were more distracted when viewing the videos. Results lend support to the effectiveness of self-disclosure as a technique that persons who stutter can use to positively influence the perceptions of listeners.

  5. A Comparison of Coverbal Gesture Use in Oral Discourse Among Speakers With Fluent and Nonfluent Aphasia

    PubMed Central

    Law, Sam-Po; Chak, Gigi Wan-Chi

    2017-01-01

    Purpose Coverbal gesture use, which is affected by the presence and degree of aphasia, can be culturally specific. The purpose of this study was to compare gesture use among Cantonese-speaking individuals: 23 neurologically healthy speakers, 23 speakers with fluent aphasia, and 21 speakers with nonfluent aphasia. Method Multimedia data of discourse samples from these speakers were extracted from the Cantonese AphasiaBank. Gestures were independently annotated on their forms and functions to determine how gesturing rate and distribution of gestures differed across speaker groups. A multiple regression was conducted to determine the most predictive variable(s) for gesture-to-word ratio. Results Although speakers with nonfluent aphasia gestured most frequently, the rate of gesture use in counterparts with fluent aphasia did not differ significantly from controls. Different patterns of gesture functions in the 3 speaker groups revealed that gesture plays a minor role in lexical retrieval whereas its role in enhancing communication dominates among the speakers with aphasia. The percentages of complete sentences and dysfluency strongly predicted the gesturing rate in aphasia. Conclusions The current results supported the sketch model of language–gesture association. The relationship between gesture production and linguistic abilities and clinical implications for gesture-based language intervention for speakers with aphasia are also discussed. PMID:28609510

  6. Accounting for the listener: comparing the production of contrastive intonation in typically-developing speakers and speakers with autism.

    PubMed

    Kaland, Constantijn; Swerts, Marc; Krahmer, Emiel

    2013-09-01

    The present research investigates what drives the prosodic marking of contrastive information. For example, a typically developing speaker of a Germanic language like Dutch generally refers to a pink car as a "PINK car" (accented words in capitals) when a previously mentioned car was red. The main question addressed in this paper is whether contrastive intonation is produced with respect to the speaker's or (also) the listener's perspective on the preceding discourse. Furthermore, this research investigates the production of contrastive intonation by typically developing speakers and speakers with autism. The latter group is investigated because people with autism are argued to have difficulties accounting for another person's mental state and exhibit difficulties in the production and perception of accentuation and pitch range. To this end, utterances with contrastive intonation are elicited from both groups and analyzed in terms of function and form of prosody using production and perception measures. Contrary to expectations, typically developing speakers and speakers with autism produce functionally similar contrastive intonation as both groups account for both their own and their listener's perspective. However, typically developing speakers use a larger pitch range and are perceived as speaking more dynamically than speakers with autism, suggesting differences in their use of prosodic form.

  7. A Comparison of Coverbal Gesture Use in Oral Discourse Among Speakers With Fluent and Nonfluent Aphasia.

    PubMed

    Kong, Anthony Pak-Hin; Law, Sam-Po; Chak, Gigi Wan-Chi

    2017-07-12

    Coverbal gesture use, which is affected by the presence and degree of aphasia, can be culturally specific. The purpose of this study was to compare gesture use among Cantonese-speaking individuals: 23 neurologically healthy speakers, 23 speakers with fluent aphasia, and 21 speakers with nonfluent aphasia. Multimedia data of discourse samples from these speakers were extracted from the Cantonese AphasiaBank. Gestures were independently annotated on their forms and functions to determine how gesturing rate and distribution of gestures differed across speaker groups. A multiple regression was conducted to determine the most predictive variable(s) for gesture-to-word ratio. Although speakers with nonfluent aphasia gestured most frequently, the rate of gesture use in counterparts with fluent aphasia did not differ significantly from controls. Different patterns of gesture functions in the 3 speaker groups revealed that gesture plays a minor role in lexical retrieval whereas its role in enhancing communication dominates among the speakers with aphasia. The percentages of complete sentences and dysfluency strongly predicted the gesturing rate in aphasia. The current results supported the sketch model of language-gesture association. The relationship between gesture production and linguistic abilities and clinical implications for gesture-based language intervention for speakers with aphasia are also discussed.

  8. Effectiveness of communication strategies for deaf or hard of hearing workers in group settings.

    PubMed

    Haynes, Scott

    2014-01-01

    In group settings, background noise and an obstructed view of the speaker are just a few of the issues that can make workplace communication difficult for an individual who is deaf or hard of hearing. Accommodation strategies such as amplification of the speaker's voice or the use of text-based alternatives exist to address these issues. However, recent studies have shown that there are still unmet needs related to workplace communication in group settings for individuals who are deaf or hard of hearing. Identify the most common strategies used by individuals who are deaf or hard of hearing to improve communication in group settings and gauge the perceived effectiveness of those strategies. An online survey was conducted with individuals who self-identified as deaf or hard of hearing. The survey presented specific communication strategies based on three functional approaches (aural/oral, text, visual). The strategies applied to both receptive and expressive communication in five different meeting types ranging in size and purpose. 161 adults (age 22-90 yrs.) with limited hearing ability completed the survey. Text-based strategies were typically the least frequently used strategies in group settings, yet they ranked high in perceived effectiveness for receptive and expressive communication. Those who used an interpreter demonstrated a strong preference for having a qualified interpreter present in the meeting rather than an interpreter acting remotely. For expressive communication, participants in general preferred to use their own voice or signing abilities and ranked those strategies as highly effective. A more accessible workplace for individuals who are deaf or hard of hearing would incorporate more ubiquitous text-based strategy options. Also, qualified interpreters, when used, should be present in the meeting for maximum effectiveness.

  9. Education and the New Technologies. Report of the WCOTP Asian and South Pacific Regional Conference (10th, Seoul, South Korea, August 7-13, 1985).

    ERIC Educational Resources Information Center

    World Confederation of Organizations of the Teaching Profession, Morges (Switzerland).

    This report on the 1985 Asian and South Pacific Regional Conference of the World Confederation of Organizations of the Teaching Profession (WCOTP) presents the full text of three addresses presented during the opening ceremony, as well as four major presentations by invited speakers and three addresses from the closing ceremony. A day of…

  10. Cultura, comunicacion e interaccion: hacia el contexto total del lenguage y el hombre hispanicos (Culture, Communication and Interaction: Towards a Global Context of the Spanish Language and Speaker)

    ERIC Educational Resources Information Center

    Poyatos, Fernando

    1975-01-01

    The new science of Proxemic Behavior (introduced by Edward T. Hall) should be included in the basic triple structure of human communicative behavior: language-paralanguage-kinesthesia. The applications of such a science are many e.g., analysis and study of the narrative character in novels. (Text is in Spanish.) (DS)

  11. Using the Longman Mini-concordancer on Tagged and Parsed Corpora, with Special Reference to Their Use as an Aid to Grammar Learning.

    ERIC Educational Resources Information Center

    Qiao, Hong Liang; Sussex, Roland

    1996-01-01

    Presents methods for using the Longman Mini-Concordancer on tagged and parsed corpora rather than plain text corpora. The article discusses several aspects with models to be applied in the classroom as an aid to grammar learning. This paper suggests exercises suitable for teaching English to both native and nonnative speakers. (13 references)…

  12. La Linguistica Aplicada a la Relacion Paradigmatica entre los Verbos "Ser" y "Estar" (Linguistics Applied to the Paradigmatic Relationship between the Verbs "Ser" and "Estar")

    ERIC Educational Resources Information Center

    Marchetti, Magda Ruggeri

    1977-01-01

    Speakers of Italian often have problems mastering Spanish because they erroneously believe its great similiarity to Italian makes it easy to learn. One of the fundamental problems is the lack of ability to choose the correct verb, "ser" or "estar," both equivalents of the Italian "essere." (Text is in Spanish.) (CFM)

  13. Combinatorics and synchronization in natural semiotics

    NASA Astrophysics Data System (ADS)

    Orsucci, Franco; Giuliani, Alessandro; Webber, Charles; Zbilut, Joseph; Fonagy, Peter; Mazza, Marianna

    2006-03-01

    In this study the derivation of an objective metrics to appreciate the degree of structuring of written and spoken texts is presented. The proposed metrics is based on the scoring of recurrences inside a text by means of the application of recurrence quantification analysis (RQA), a nonlinear technique widely used in other fields of sciences. The adopted approach allowed us to create a ranking of different poems strictly related to their prosodic structure and, more importantly, the possibility to recognize the same structure across different languages, to define a level of structuring typical of spoken texts and identifying the progressive synchronization of a dyadic relation between two speakers in terms of relative complexity of their speeches. These results suggest the possibility of introducing objective measurement methods into humanities studies.

  14. Shhh… I Need Quiet! Children's Understanding of American, British, and Japanese-accented English Speakers.

    PubMed

    Bent, Tessa; Holt, Rachael Frush

    2018-02-01

    Children's ability to understand speakers with a wide range of dialects and accents is essential for efficient language development and communication in a global society. Here, the impact of regional dialect and foreign-accent variability on children's speech understanding was evaluated in both quiet and noisy conditions. Five- to seven-year-old children ( n = 90) and adults ( n = 96) repeated sentences produced by three speakers with different accents-American English, British English, and Japanese-accented English-in quiet or noisy conditions. Adults had no difficulty understanding any speaker in quiet conditions. Their performance declined for the nonnative speaker with a moderate amount of noise; their performance only substantially declined for the British English speaker (i.e., below 93% correct) when their understanding of the American English speaker was also impeded. In contrast, although children showed accurate word recognition for the American and British English speakers in quiet conditions, they had difficulty understanding the nonnative speaker even under ideal listening conditions. With a moderate amount of noise, their perception of British English speech declined substantially and their ability to understand the nonnative speaker was particularly poor. These results suggest that although school-aged children can understand unfamiliar native dialects under ideal listening conditions, their ability to recognize words in these dialects may be highly susceptible to the influence of environmental degradation. Fully adult-like word identification for speakers with unfamiliar accents and dialects may exhibit a protracted developmental trajectory.

  15. How Do Speakers Avoid Ambiguous Linguistic Expressions?

    ERIC Educational Resources Information Center

    Ferreira, V.S.; Slevc, L.R.; Rogers, E.S.

    2005-01-01

    Three experiments assessed how speakers avoid linguistically and nonlinguistically ambiguous expressions. Speakers described target objects (a flying mammal, bat) in contexts including foil objects that caused linguistic (a baseball bat) and nonlinguistic (a larger flying mammal) ambiguity. Speakers sometimes avoided linguistic-ambiguity, and they…

  16. Combining Behavioral and ERP Methodologies to Investigate the Differences Between McGurk Effects Demonstrated by Cantonese and Mandarin Speakers.

    PubMed

    Zhang, Juan; Meng, Yaxuan; McBride, Catherine; Fan, Xitao; Yuan, Zhen

    2018-01-01

    The present study investigated the impact of Chinese dialects on McGurk effect using behavioral and event-related potential (ERP) methodologies. Specifically, intra-language comparison of McGurk effect was conducted between Mandarin and Cantonese speakers. The behavioral results showed that Cantonese speakers exhibited a stronger McGurk effect in audiovisual speech perception compared to Mandarin speakers, although both groups performed equally in the auditory and visual conditions. ERP results revealed that Cantonese speakers were more sensitive to visual cues than Mandarin speakers, though this was not the case for the auditory cues. Taken together, the current findings suggest that the McGurk effect generated by Chinese speakers is mainly influenced by segmental phonology during audiovisual speech integration.

  17. Combining Behavioral and ERP Methodologies to Investigate the Differences Between McGurk Effects Demonstrated by Cantonese and Mandarin Speakers

    PubMed Central

    Zhang, Juan; Meng, Yaxuan; McBride, Catherine; Fan, Xitao; Yuan, Zhen

    2018-01-01

    The present study investigated the impact of Chinese dialects on McGurk effect using behavioral and event-related potential (ERP) methodologies. Specifically, intra-language comparison of McGurk effect was conducted between Mandarin and Cantonese speakers. The behavioral results showed that Cantonese speakers exhibited a stronger McGurk effect in audiovisual speech perception compared to Mandarin speakers, although both groups performed equally in the auditory and visual conditions. ERP results revealed that Cantonese speakers were more sensitive to visual cues than Mandarin speakers, though this was not the case for the auditory cues. Taken together, the current findings suggest that the McGurk effect generated by Chinese speakers is mainly influenced by segmental phonology during audiovisual speech integration. PMID:29780312

  18. Left-lateralized N170 Effects of Visual Expertise in Reading: Evidence from Japanese Syllabic and Logographic Scripts

    PubMed Central

    Maurer, Urs; Zevin, Jason D.; McCandliss, Bruce D.

    2015-01-01

    The N170 component of the event-related potential (ERP) reflects experience-dependent neural changes in several forms of visual expertise, including expertise for visual words. Readers skilled in writing systems that link characters to phonemes (i.e., alphabetic writing) typically produce a left-lateralized N170 to visual word forms. This study examined the N170 in three Japanese scripts that link characters to larger phonological units. Participants were monolingual English speakers (EL1) and native Japanese speakers (JL1) who were also proficient in English. ERPs were collected using a 129-channel array, as participants performed a series of experiments viewing words or novel control stimuli in a repetition detection task. The N170 was strongly left-lateralized for all three Japanese scripts (including logographic Kanji characters) in JL1 participants, but bilateral in EL1 participants viewing these same stimuli. This demonstrates that left-lateralization of the N170 is dependent on specific reading expertise and is not limited to alphabetic scripts. Additional contrasts within the moraic Katakana script revealed equivalent N170 responses in JL1 speakers for familiar Katakana words and for Kanji words transcribed into novel Katakana words, suggesting that the N170 expertise effect is driven by script familiarity rather than familiarity with particular visual word forms. Finally, for English words and novel symbol string stimuli, both EL1 and JL1 subjects produced equivalent responses for the novel symbols, and more left-lateralized N170 responses for the English words, indicating that such effects are not limited to the first language. Taken together, these cross-linguistic results suggest that similar neural processes underlie visual expertise for print in very different writing systems. PMID:18370600

  19. Objective eye-gaze behaviour during face-to-face communication with proficient alaryngeal speakers: a preliminary study.

    PubMed

    Evitts, Paul; Gallop, Robert

    2011-01-01

    There is a large body of research demonstrating the impact of visual information on speaker intelligibility in both normal and disordered speaker populations. However, there is minimal information on which specific visual features listeners find salient during conversational discourse. To investigate listeners' eye-gaze behaviour during face-to-face conversation with normal, laryngeal and proficient alaryngeal speakers. Sixty participants individually participated in a 10-min conversation with one of four speakers (typical laryngeal, tracheoesophageal, oesophageal, electrolaryngeal; 15 participants randomly assigned to one mode of speech). All speakers were > 85% intelligible and were judged to be 'proficient' by two certified speech-language pathologists. Participants were fitted with a head-mounted eye-gaze tracking device (Mobile Eye, ASL) that calculated the region of interest and mean duration of eye-gaze. Self-reported gaze behaviour was also obtained following the conversation using a 10 cm visual analogue scale. While listening, participants viewed the lower facial region of the oesophageal speaker more than the normal or tracheoesophageal speaker. Results of non-hierarchical cluster analyses showed that while listening, the pattern of eye-gaze was predominantly directed at the lower face of the oesophageal and electrolaryngeal speaker and more evenly dispersed among the background, lower face, and eyes of the normal and tracheoesophageal speakers. Finally, results show a low correlation between self-reported eye-gaze behaviour and objective regions of interest data. Overall, results suggest similar eye-gaze behaviour when healthy controls converse with normal and tracheoesophageal speakers and that participants had significantly different eye-gaze patterns when conversing with an oesophageal speaker. Results are discussed in terms of existing eye-gaze data and its potential implications on auditory-visual speech perception. © 2011 Royal College of Speech & Language Therapists.

  20. Speech serial control in healthy speakers and speakers with hypokinetic or ataxic dysarthria: effects of sequence length and practice

    PubMed Central

    Reilly, Kevin J.; Spencer, Kristie A.

    2013-01-01

    The current study investigated the processes responsible for selection of sounds and syllables during production of speech sequences in 10 adults with hypokinetic dysarthria from Parkinson’s disease, five adults with ataxic dysarthria, and 14 healthy control speakers. Speech production data from a choice reaction time task were analyzed to evaluate the effects of sequence length and practice on speech sound sequencing. Speakers produced sequences that were between one and five syllables in length over five experimental runs of 60 trials each. In contrast to the healthy speakers, speakers with hypokinetic dysarthria demonstrated exaggerated sequence length effects for both inter-syllable intervals (ISIs) and speech error rates. Conversely, speakers with ataxic dysarthria failed to demonstrate a sequence length effect on ISIs and were also the only group that did not exhibit practice-related changes in ISIs and speech error rates over the five experimental runs. The exaggerated sequence length effects in the hypokinetic speakers with Parkinson’s disease are consistent with an impairment of action selection during speech sequence production. The absent length effects observed in the speakers with ataxic dysarthria is consistent with previous findings that indicate a limited capacity to buffer speech sequences in advance of their execution. In addition, the lack of practice effects in these speakers suggests that learning-related improvements in the production rate and accuracy of speech sequences involves processing by structures of the cerebellum. Together, the current findings inform models of serial control for speech in healthy speakers and support the notion that sequencing deficits contribute to speech symptoms in speakers with hypokinetic or ataxic dysarthria. In addition, these findings indicate that speech sequencing is differentially impaired in hypokinetic and ataxic dysarthria. PMID:24137121

  1. Discourse comprehension in L2: Making sense of what is not explicitly said.

    PubMed

    Foucart, Alice; Romero-Rivas, Carlos; Gort, Bernharda Lottie; Costa, Albert

    2016-12-01

    Using ERPs, we tested whether L2 speakers can integrate multiple sources of information (e.g., semantic, pragmatic information) during discourse comprehension. We presented native speakers and L2 speakers with three-sentence scenarios in which the final sentence was highly causally related, intermediately related, or causally unrelated to its context; its interpretation therefore required simple or complex inferences. Native speakers revealed a gradual N400-like effect, larger in the causally unrelated condition than in the highly related condition, and falling in-between in the intermediately related condition, replicating previous results. In the crucial intermediately related condition, L2 speakers behaved like native speakers, however, showing extra processing in a later time-window. Overall, the results show that, when reading, L2 speakers are able to process information from the local context and prior information (e.g., world knowledge) to build global coherence, suggesting that they process different sources of information to make inferences online during discourse comprehension, like native speakers. Copyright © 2016 Elsevier Inc. All rights reserved.

  2. Speaker recognition with temporal cues in acoustic and electric hearing

    NASA Astrophysics Data System (ADS)

    Vongphoe, Michael; Zeng, Fan-Gang

    2005-08-01

    Natural spoken language processing includes not only speech recognition but also identification of the speaker's gender, age, emotional, and social status. Our purpose in this study is to evaluate whether temporal cues are sufficient to support both speech and speaker recognition. Ten cochlear-implant and six normal-hearing subjects were presented with vowel tokens spoken by three men, three women, two boys, and two girls. In one condition, the subject was asked to recognize the vowel. In the other condition, the subject was asked to identify the speaker. Extensive training was provided for the speaker recognition task. Normal-hearing subjects achieved nearly perfect performance in both tasks. Cochlear-implant subjects achieved good performance in vowel recognition but poor performance in speaker recognition. The level of the cochlear implant performance was functionally equivalent to normal performance with eight spectral bands for vowel recognition but only to one band for speaker recognition. These results show a disassociation between speech and speaker recognition with primarily temporal cues, highlighting the limitation of current speech processing strategies in cochlear implants. Several methods, including explicit encoding of fundamental frequency and frequency modulation, are proposed to improve speaker recognition for current cochlear implant users.

  3. Development of a speaker discrimination test for cochlear implant users based on the Oldenburg Logatome corpus.

    PubMed

    Mühler, Roland; Ziese, Michael; Rostalski, Dorothea

    2009-01-01

    The purpose of the study was to develop a speaker discrimination test for cochlear implant (CI) users. The speech material was drawn from the Oldenburg Logatome (OLLO) corpus, which contains 150 different logatomes read by 40 German and 10 French native speakers. The prototype test battery included 120 logatome pairs spoken by 5 male and 5 female speakers with balanced representations of the conditions 'same speaker' and 'different speaker'. Ten adult normal-hearing listeners and 12 adult postlingually deafened CI users were included in a study to evaluate the suitability of the test. The mean speaker discrimination score for the CI users was 67.3% correct and for the normal-hearing listeners 92.2% correct. A significant influence of voice gender and fundamental frequency difference on the speaker discrimination score was found in CI users as well as in normal-hearing listeners. Since the test results of the CI users were significantly above chance level and no ceiling effect was observed, we conclude that subsets of the OLLO corpus are very well suited to speaker discrimination experiments in CI users. Copyright 2008 S. Karger AG, Basel.

  4. Speaker Clustering for a Mixture of Singing and Reading (Preprint)

    DTIC Science & Technology

    2012-03-01

    diarization [2, 3] which answers the ques- tion of ”who spoke when?” is a combination of speaker segmentation and clustering. Although it is possible to...focuses on speaker clustering, the techniques developed here can be applied to speaker diarization . For the remainder of this paper, the term ”speech...and retrieval,” Proceedings of the IEEE, vol. 88, 2000. [2] S. Tranter and D. Reynolds, “An overview of automatic speaker diarization systems,” IEEE

  5. Learning Words from Speakers with False Beliefs

    ERIC Educational Resources Information Center

    Papafragou, Anna; Fairchild, Sarah; Cohen, Matthew L.; Friedberg, Carlyn

    2017-01-01

    During communication, hearers try to infer the speaker's intentions to be able to understand what the speaker means. Nevertheless, whether (and how early) preschoolers track their interlocutors' mental states is still a matter of debate. Furthermore, there is disagreement about how children's ability to consult a speaker's belief in communicative…

  6. International Student Speaker Programs: "Someone from Another World."

    ERIC Educational Resources Information Center

    Wilson, Angene

    This study surveyed members of the Association of International Educators and community volunteers to find out how international student speaker programs actually work. An international student speaker program provides speakers (from the university foreign student population) for community organizations and schools. The results of the survey (49…

  7. Linguistic "Mudes" and the De-Ethnicization of Language Choice in Catalonia

    ERIC Educational Resources Information Center

    Pujolar, Joan; Gonzalez, Isaac

    2013-01-01

    Catalan speakers have traditionally constructed the Catalan language as the main emblem of their identity even as migration filled the country with substantial numbers of speakers of Castilian. Although Catalan speakers have been bilingual in Catalan and Castilian for generations, sociolinguistic research has shown how speakers' bilingual…

  8. Embodied Communication: Speakers' Gestures Affect Listeners' Actions

    ERIC Educational Resources Information Center

    Cook, Susan Wagner; Tanenhaus, Michael K.

    2009-01-01

    We explored how speakers and listeners use hand gestures as a source of perceptual-motor information during naturalistic communication. After solving the Tower of Hanoi task either with real objects or on a computer, speakers explained the task to listeners. Speakers' hand gestures, but not their speech, reflected properties of the particular…

  9. Speech Breathing in Speakers Who Use an Electrolarynx

    ERIC Educational Resources Information Center

    Bohnenkamp, Todd A.; Stowell, Talena; Hesse, Joy; Wright, Simon

    2010-01-01

    Speakers who use an electrolarynx following a total laryngectomy no longer require pulmonary support for speech. Subsequently, chest wall movements may be affected; however, chest wall movements in these speakers are not well defined. The purpose of this investigation was to evaluate speech breathing in speakers who use an electrolarynx during…

  10. A multi-views multi-learners approach towards dysarthric speech recognition using multi-nets artificial neural networks.

    PubMed

    Shahamiri, Seyed Reza; Salim, Siti Salwah Binti

    2014-09-01

    Automatic speech recognition (ASR) can be very helpful for speakers who suffer from dysarthria, a neurological disability that damages the control of motor speech articulators. Although a few attempts have been made to apply ASR technologies to sufferers of dysarthria, previous studies show that such ASR systems have not attained an adequate level of performance. In this study, a dysarthric multi-networks speech recognizer (DM-NSR) model is provided using a realization of multi-views multi-learners approach called multi-nets artificial neural networks, which tolerates variability of dysarthric speech. In particular, the DM-NSR model employs several ANNs (as learners) to approximate the likelihood of ASR vocabulary words and to deal with the complexity of dysarthric speech. The proposed DM-NSR approach was presented as both speaker-dependent and speaker-independent paradigms. In order to highlight the performance of the proposed model over legacy models, multi-views single-learner models of the DM-NSRs were also provided and their efficiencies were compared in detail. Moreover, a comparison among the prominent dysarthric ASR methods and the proposed one is provided. The results show that the DM-NSR recorded improved recognition rate by up to 24.67% and the error rate was reduced by up to 8.63% over the reference model.

  11. Auditory evoked fields to vocalization during passive listening and active generation in adults who stutter.

    PubMed

    Beal, Deryk S; Cheyne, Douglas O; Gracco, Vincent L; Quraan, Maher A; Taylor, Margot J; De Nil, Luc F

    2010-10-01

    We used magnetoencephalography to investigate auditory evoked responses to speech vocalizations and non-speech tones in adults who do and do not stutter. Neuromagnetic field patterns were recorded as participants listened to a 1 kHz tone, playback of their own productions of the vowel /i/ and vowel-initial words, and actively generated the vowel /i/ and vowel-initial words. Activation of the auditory cortex at approximately 50 and 100 ms was observed during all tasks. A reduction in the peak amplitudes of the M50 and M100 components was observed during the active generation versus passive listening tasks dependent on the stimuli. Adults who stutter did not differ in the amount of speech-induced auditory suppression relative to fluent speakers. Adults who stutter had shorter M100 latencies for the actively generated speaking tasks in the right hemisphere relative to the left hemisphere but the fluent speakers showed similar latencies across hemispheres. During passive listening tasks, adults who stutter had longer M50 and M100 latencies than fluent speakers. The results suggest that there are timing, rather than amplitude, differences in auditory processing during speech in adults who stutter and are discussed in relation to hypotheses of auditory-motor integration breakdown in stuttering. Copyright 2010 Elsevier Inc. All rights reserved.

  12. Top–Down Modulation on the Perception and Categorization of Identical Pitch Contours in Speech and Music

    PubMed Central

    Weidema, Joey L.; Roncaglia-Denissen, M. P.; Honing, Henkjan

    2016-01-01

    Whether pitch in language and music is governed by domain-specific or domain-general cognitive mechanisms is contentiously debated. The aim of the present study was to investigate whether mechanisms governing pitch contour perception operate differently when pitch information is interpreted as either speech or music. By modulating listening mode, this study aspired to demonstrate that pitch contour perception relies on domain-specific cognitive mechanisms, which are regulated by top–down influences from language and music. Three groups of participants (Mandarin speakers, Dutch speaking non-musicians, and Dutch musicians) were exposed to identical pitch contours, and tested on their ability to identify these contours in a language and musical context. Stimuli consisted of disyllabic words spoken in Mandarin, and melodic tonal analogs, embedded in a linguistic and melodic carrier phrase, respectively. Participants classified identical pitch contours as significantly different depending on listening mode. Top–down influences from language appeared to alter the perception of pitch contour in speakers of Mandarin. This was not the case for non-musician speakers of Dutch. Moreover, this effect was lacking in Dutch speaking musicians. The classification patterns of pitch contours in language and music seem to suggest that domain-specific categorization is modulated by top–down influences from language and music. PMID:27313552

  13. Speaker-independent factors affecting the perception of foreign accent in a second languagea)

    PubMed Central

    Levi, Susannah V.; Winters, Stephen J.; Pisoni, David B.

    2012-01-01

    Previous research on foreign accent perception has largely focused on speaker-dependent factors such as age of learning and length of residence. Factors that are independent of a speaker’s language learning history have also been shown to affect perception of second language speech. The present study examined the effects of two such factors—listening context and lexical frequency—on the perception of foreign-accented speech. Listeners rated foreign accent in two listening contexts: auditory-only, where listeners only heard the target stimuli, and auditory+orthography, where listeners were presented with both an auditory signal and an orthographic display of the target word. Results revealed that higher frequency words were consistently rated as less accented than lower frequency words. The effect of the listening context emerged in two interactions: the auditory +orthography context reduced the effects of lexical frequency, but increased the perceived differences between native and non-native speakers. Acoustic measurements revealed some production differences for words of different levels of lexical frequency, though these differences could not account for all of the observed interactions from the perceptual experiment. These results suggest that factors independent of the speakers’ actual speech articulations can influence the perception of degree of foreign accent. PMID:17471745

  14. 2016 Microbial Stress Response GRC/GRS

    DTIC Science & Technology

    2016-09-13

    Holyoke College South Hadley, MA Chairs: Eduardo A. Groisman & Dianne K. Newman Vice Chairs: Petra A. Levin & William W. Navarre Contributors...by Discussion Leader 9:10 am - 9:35 am Martin Ackermann (ETH Zurich, Switzerland) "History-Dependence in Bacterial Stress Response – Scaling up from...Government. Microbial Stress Response GRC – Registration List Ackermann, Martin ETH Zurich Speaker Registered Andersson, Dan I Uppsala

  15. STS-41 Voice Command System Flight Experiment Report

    NASA Technical Reports Server (NTRS)

    Salazar, George A.

    1981-01-01

    This report presents the results of the Voice Command System (VCS) flight experiment on the five-day STS-41 mission. Two mission specialists,Bill Shepherd and Bruce Melnick, used the speaker-dependent system to evaluate the operational effectiveness of using voice to control a spacecraft system. In addition, data was gathered to analyze the effects of microgravity on speech recognition performance.

  16. Processing Subject-Verb Agreement in a Second Language Depends on Proficiency

    ERIC Educational Resources Information Center

    Hoshino, Noriko; Dussias, Paola E.; Kroll, Judith F.

    2010-01-01

    Subject-verb agreement is a computation that is often difficult to execute perfectly in the first language (L1) and even more difficult to produce skillfully in a second language (L2). In this study, we examine the way in which bilingual speakers complete sentence fragments in a manner that reflects access to both grammatical and conceptual…

  17. Second Language Processing: When Are First and Second Languages Processed Similarly?

    ERIC Educational Resources Information Center

    Sabourin, Laura; Stowe, Laurie A.

    2008-01-01

    In this article we investigate the effects of first language (L1) on second language (L2) neural processing for two grammatical constructions (verbal domain dependency and grammatical gender), focusing on the event-related potential P600 effect, which has been found in both L1 and L2 processing. Native Dutch speakers showed a P600 effect for both…

  18. Processing of Tense Morphology and Filler-Gap Dependencies by Chinese Second Language Speakers of English

    ERIC Educational Resources Information Center

    Dong, Zhiyin Renee

    2014-01-01

    There is an ongoing debate in the field of Second Language Acquisition concerning whether a fundamental difference exists between the native language (L1) and adult second language (L2) online processing of syntax and morpho-syntax. The Shallow Structure Hypothesis (SSH) (Clahsen and Felser, 2006a, b) states that L2 online parsing is qualitatively…

  19. What Do You Teach...? The Role of Argument in Rhetorical Invention: An Integrated Skills Approach.

    ERIC Educational Resources Information Center

    Stelzner, Sara L.

    Speaking and writing should be taught together as they both are concerned with the communication model that includes a speaker, a listener, and a subject and the way these elements affect each other. In speaking, it is clear that invention is a public process depending on the listener's or receiver's active participation in the creation of…

  20. Timed and Untimed Grammaticality Judgments Measure Distinct Types of Knowledge: Evidence from Eye-Movement Patterns

    ERIC Educational Resources Information Center

    Godfroid, Aline; Loewen, Shawn; Jung, Sehoon; Park, Ji-Hyun; Gass, Susan; Ellis, Rod

    2015-01-01

    Grammaticality judgment tests (GJTs) have been used to elicit data reflecting second language (L2) speakers' knowledge of L2 grammar. However, the exact constructs measured by GJTs, whether primarily implicit or explicit knowledge, are disputed and have been argued to differ depending on test-related variables (i.e., time pressure and item…

  1. Text Message Support for Weight Loss in Patients With Prediabetes: A Randomized Clinical Trial.

    PubMed

    Fischer, Henry H; Fischer, Ilana P; Pereira, Rocio I; Furniss, Anna L; Rozwadowski, Jeanne M; Moore, Susan L; Durfee, Michael J; Raghunath, Silvia G; Tsai, Adam G; Havranek, Edward P

    2016-08-01

    Although the benefits of in-person Diabetes Prevention Program (DPP) classes for diabetes prevention have been demonstrated in trials, effectiveness in clinical practice is limited by low participation rates. This study explores whether text message support enhances weight loss in patients offered DPP classes. English- and Spanish-speaking patients with prediabetes (n = 163) were randomized to the control group, which only received an invitation to DPP classes as defined by the Centers for Disease Control and Prevention, or to the text message-augmented intervention group, which also received text messages adapted from the DPP curriculum for 12 months. Mean weight decreased 0.6 pounds (95% CI -2.7 to 1.6) in the control group and 2.6 pounds (95% CI -5.5 to 0.2) in the intervention group (P value 0.05). Three percent weight loss was achieved by 21.5% of participants in the control group (95% CI 12.5-30.6), compared with 38.5% in the intervention group (95% CI 27.7-49.3) (absolute difference 17.0%; P value 0.02). Mean glycated hemoglobin (HbA1c) increased by 0.19% or 2.1 mmol/mol (95% CI -0.1 to 0.5%) and decreased by 0.09% or 1.0 mmol/mol (95% CI -0.2 to 0.0%) in the control group and intervention participants, respectively (absolute difference 0.28%; P value 0.07). Stratification by language demonstrated a significant treatment effect in Spanish speakers but not in English speakers. Text message support can lead to clinically significant weight loss in patients with prediabetes. Further study assessing effect by primary language and in an operational setting is warranted. © 2016 by the American Diabetes Association. Readers may use this article as long as the work is properly cited, the use is educational and not for profit, and the work is not altered.

  2. The speakers' bureau system: a form of peer selling.

    PubMed

    Reid, Lynette; Herder, Matthew

    2013-01-01

    In the speakers' bureau system, physicians are recruited and trained by pharmaceutical, biotechnology, and medical device companies to deliver information about products to other physicians, in exchange for a fee. Using publicly available disclosures, we assessed the thesis that speakers' bureau involvement is not a feature of academic medicine in Canada, by estimating the prevalence of participation in speakers' bureaus among Canadian faculty in one medical specialty, cardiology. We analyzed the relevant features of an actual contract made public by the physician addressee and applied the Canadian Medical Association (CMA) guidelines on physician-industry relations to participation in a speakers' bureau. We argue that speakers' bureau participation constitutes a form of peer selling that should be understood to contravene the prohibition on product endorsement in the CMA Code of Ethics. Academic medical institutions, in conjunction with regulatory colleges, should continue and strengthen their policies to address participation in speakers' bureaus.

  3. A Syntactic Account for the Power of Verbs within X-Phemism: A Corpus-Base Exploration

    ERIC Educational Resources Information Center

    Al-Rawi, Maather Mohammad; Al-Shurafa, Nuha Suleiman Daoud

    2016-01-01

    The main aim of this paper is to examine the syntactic status of a selected text-corpus focus, with a special focus on the verb within its Verb-Phrase. The major claim is that the power of the verb in its VP is loaded syntactically through which the speaker's desire of the doublespeak within X-Phemism is achieved. In order to fulfill this claim, a…

  4. Simultaneous Talk--From the Perspective of Floor Management of English and Japanese Speakers.

    ERIC Educational Resources Information Center

    Hayashi, Reiko

    1988-01-01

    Investigates simultaneous talk in face-to-face conversation using the analytic framework of "floor" proposed by Edelsky (1981). Analysis of taped conversation among speakers of Japanese and among speakers of English shows that, while both groups use simultaneous talk, it is used more frequently by Japanese speakers. A reference list…

  5. Respiratory Control in Stuttering Speakers: Evidence from Respiratory High-Frequency Oscillations.

    ERIC Educational Resources Information Center

    Denny, Margaret; Smith, Anne

    2000-01-01

    This study examined whether stuttering speakers (N=10) differed from fluent speakers in relations between the neural control systems for speech and life support. It concluded that in some stuttering speakers the relations between respiratory controllers are atypical, but that high participation by the high frequency oscillation-producing circuitry…

  6. The Effects of Source Unreliability on Prior and Future Word Learning

    ERIC Educational Resources Information Center

    Faught, Gayle G.; Leslie, Alicia D.; Scofield, Jason

    2015-01-01

    Young children regularly learn words from interactions with other speakers, though not all speakers are reliable informants. Interestingly, children will reverse to trusting a reliable speaker when a previously endorsed speaker proves unreliable. When later asked to identify the referent of a novel word, children who reverse trust are less willing…

  7. The Semantic Basis of Do So.

    ERIC Educational Resources Information Center

    Binder, Richard

    The thesis of this paper is that the "do so" test described by Lakoff and Ross (1966) is a test of the speaker's belief system regarding the relationship of verbs to their surface subject, and that judgments of grammaticality concerning "do so" are based on the speaker's underlying semantic beliefs. ("Speaker" refers here to both speakers and…

  8. Speaker Reliability Guides Children's Inductive Inferences about Novel Properties

    ERIC Educational Resources Information Center

    Kim, Sunae; Kalish, Charles W.; Harris, Paul L.

    2012-01-01

    Prior work shows that children can make inductive inferences about objects based on their labels rather than their appearance (Gelman, 2003). A separate line of research shows that children's trust in a speaker's label is selective. Children accept labels from a reliable speaker over an unreliable speaker (e.g., Koenig & Harris, 2005). In the…

  9. Native-Speakerism and the Complexity of Personal Experience: A Duoethnographic Study

    ERIC Educational Resources Information Center

    Lowe, Robert J.; Kiczkowiak, Marek

    2016-01-01

    This paper presents a duoethnographic study into the effects of native-speakerism on the professional lives of two English language teachers, one "native", and one "non-native speaker" of English. The goal of the study was to build on and extend existing research on the topic of native-speakerism by investigating, through…

  10. Research Timeline: Second Language Communication Strategies

    ERIC Educational Resources Information Center

    Kennedy, Sara; Trofimovich, Pavel

    2016-01-01

    Speakers of a second language (L2), regardless of profciency level, communicate for specifc purposes. For example, an L2 speaker of English may wish to build rapport with a co-worker by chatting about the weather. The speaker will draw on various resources to accomplish her communicative purposes. For instance, the speaker may say "falling…

  11. Word Stress and Pronunciation Teaching in English as a Lingua Franca Contexts

    ERIC Educational Resources Information Center

    Lewis, Christine; Deterding, David

    2018-01-01

    Traditionally, pronunciation was taught by reference to native-speaker models. However, as speakers around the world increasingly interact in English as a lingua franca (ELF) contexts, there is less focus on native-speaker targets, and there is wide acceptance that achieving intelligibility is crucial while mimicking native-speaker pronunciation…

  12. Defining "Native Speaker" in Multilingual Settings: English as a Native Language in Asia

    ERIC Educational Resources Information Center

    Hansen Edwards, Jette G.

    2017-01-01

    The current study examines how and why speakers of English from multilingual contexts in Asia are identifying as native speakers of English. Eighteen participants from different contexts in Asia, including Singapore, Malaysia, India, Taiwan, and The Philippines, who self-identified as native speakers of English participated in hour-long interviews…

  13. Speaker Identity Supports Phonetic Category Learning

    ERIC Educational Resources Information Center

    Mani, Nivedita; Schneider, Signe

    2013-01-01

    Visual cues from the speaker's face, such as the discriminable mouth movements used to produce speech sounds, improve discrimination of these sounds by adults. The speaker's face, however, provides more information than just the mouth movements used to produce speech--it also provides a visual indexical cue of the identity of the speaker. The…

  14. The Interpretability Hypothesis: Evidence from Wh-Interrogatives in Second Language Acquisition

    ERIC Educational Resources Information Center

    Tsimpli, Ianthi Maria; Dimitrakopoulou, Maria

    2007-01-01

    The second language acquisition (SLA) literature reports numerous studies of proficient second language (L2) speakers who diverge significantly from native speakers despite the evidence offered by the L2 input. Recent SLA theories have attempted to account for native speaker/non-native speaker (NS/NNS) divergence by arguing for the dissociation…

  15. The interaction of glottal-pulse rate and vocal-tract length in judgements of speaker size, sex, and age

    NASA Astrophysics Data System (ADS)

    Smith, David R. R.; Patterson, Roy D.

    2005-11-01

    Glottal-pulse rate (GPR) and vocal-tract length (VTL) are related to the size, sex, and age of the speaker but it is not clear how the two factors combine to influence our perception of speaker size, sex, and age. This paper describes experiments designed to measure the effect of the interaction of GPR and VTL upon judgements of speaker size, sex, and age. Vowels were scaled to represent people with a wide range of GPRs and VTLs, including many well beyond the normal range of the population, and listeners were asked to judge the size and sex/age of the speaker. The judgements of speaker size show that VTL has a strong influence upon perceived speaker size. The results for the sex and age categorization (man, woman, boy, or girl) show that, for vowels with GPR and VTL values in the normal range, judgements of speaker sex and age are influenced about equally by GPR and VTL. For vowels with abnormal combinations of low GPRs and short VTLs, the VTL information appears to decide the sex/age judgement.

  16. Comparison of singer's formant, speaker's ring, and LTA spectrum among classical singers and untrained normal speakers.

    PubMed

    Oliveira Barrichelo, V M; Heuer, R J; Dean, C M; Sataloff, R T

    2001-09-01

    Many studies have described and analyzed the singer's formant. A similar phenomenon produced by trained speakers led some authors to examine the speaker's ring. If we consider these phenomena as resonance effects associated with vocal tract adjustments and training, can we hypothesize that trained singers can carry over their singing formant ability into speech, also obtaining a speaker's ring? Can we find similar differences for energy distribution in continuous speech? Forty classically trained singers and forty untrained normal speakers performed an all-voiced reading task and produced a sample of a sustained spoken vowel /a/. The singers were also requested to perform a sustained sung vowel /a/ at a comfortable pitch. The reading was analyzed by the long-term average spectrum (LTAS) method. The sustained vowels were analyzed through power spectrum analysis. The data suggest that singers show more energy concentration in the singer's formant/speaker's ring region in both sung and spoken vowels. The singers' spoken vowel energy in the speaker's ring area was found to be significantly larger than that of the untrained speakers. The LTAS showed similar findings suggesting that those differences also occur in continuous speech. This finding supports the value of further research on the effect of singing training on the resonance of the speaking voice.

  17. Talker and accent variability effects on spoken word recognition

    NASA Astrophysics Data System (ADS)

    Nyang, Edna E.; Rogers, Catherine L.; Nishi, Kanae

    2003-04-01

    A number of studies have shown that words in a list are recognized less accurately in noise and with longer response latencies when they are spoken by multiple talkers, rather than a single talker. These results have been interpreted as support for an exemplar-based model of speech perception, in which it is assumed that detailed information regarding the speaker's voice is preserved in memory and used in recognition, rather than being eliminated via normalization. In the present study, the effects of varying both accent and talker are investigated using lists of words spoken by (a) a single native English speaker, (b) six native English speakers, (c) three native English speakers and three Japanese-accented English speakers. Twelve /hVd/ words were mixed with multi-speaker babble at three signal-to-noise ratios (+10, +5, and 0 dB) to create the word lists. Native English-speaking listeners' percent-correct recognition for words produced by native English speakers across the three talker conditions (single talker native, multi-talker native, and multi-talker mixed native and non-native) and three signal-to-noise ratios will be compared to determine whether sources of speaker variability other than voice alone add to the processing demands imposed by simple (i.e., single accent) speaker variability in spoken word recognition.

  18. Speaker and Observer Perceptions of Physical Tension during Stuttering.

    PubMed

    Tichenor, Seth; Leslie, Paula; Shaiman, Susan; Yaruss, J Scott

    2017-01-01

    Speech-language pathologists routinely assess physical tension during evaluation of those who stutter. If speakers experience tension that is not visible to clinicians, then judgments of severity may be inaccurate. This study addressed this potential discrepancy by comparing judgments of tension by people who stutter and expert clinicians to determine if clinicians could accurately identify the speakers' experience of physical tension. Ten adults who stutter were audio-video recorded in two speaking samples. Two board-certified specialists in fluency evaluated the samples using the Stuttering Severity Instrument-4 and a checklist adapted for this study. Speakers rated their tension using the same forms, and then discussed their experiences in a qualitative interview so that themes related to physical tension could be identified. The degree of tension reported by speakers was higher than that observed by specialists. Tension in parts of the body that were less visible to the observer (chest, abdomen, throat) was reported more by speakers than by specialists. The thematic analysis revealed that speakers' experience of tension changes over time and that these changes may be related to speakers' acceptance of stuttering. The lack of agreement between speaker and specialist perceptions of tension suggests that using self-reports is a necessary component for supporting the accurate diagnosis of tension in stuttering. © 2018 S. Karger AG, Basel.

  19. Non-English speakers attend gastroenterology clinic appointments at higher rates than English speakers in a vulnerable patient population

    PubMed Central

    Sewell, Justin L.; Kushel, Margot B.; Inadomi, John M.; Yee, Hal F.

    2009-01-01

    Goals We sought to identify factors associated with gastroenterology clinic attendance in an urban safety net healthcare system. Background Missed clinic appointments reduce the efficiency and availability of healthcare, but subspecialty clinic attendance among patients with established healthcare access has not been studied. Study We performed an observational study using secondary data from administrative sources to study patients referred to, and scheduled for an appointment in, the adult gastroenterology clinic serving the safety net healthcare system of San Francisco, California. Our dependent variable was whether subjects attended or missed a scheduled appointment. Analysis included multivariable logistic regression and classification tree analysis. 1,833 patients were referred and scheduled for an appointment between 05/2005 and 08/2006. Prisoners were excluded. All patients had a primary care provider. Results 683 patients (37.3%) missed their appointment; 1,150 (62.7%) attended. Language was highly associated with attendance in the logistic regression; non-English speakers were less likely than English speakers to miss an appointment (adjusted odds ratio 0.42 [0.28,0.63] for Spanish, 0.56 [0.38,0.82] for Asian language, p < 0.001). Other factors were also associated with attendance, but classification tree analysis identified language to be the most highly associated variable. Conclusions In an urban safety net healthcare population, among patients with established healthcare access and a scheduled gastroenterology clinic appointment, not speaking English was most strongly associated with higher attendance rates. Patient related factors associated with not speaking English likely influence subspecialty clinic attendance rates, and these factors may differ from those affecting general healthcare access. PMID:19169147

  20. Performance enhancement for audio-visual speaker identification using dynamic facial muscle model.

    PubMed

    Asadpour, Vahid; Towhidkhah, Farzad; Homayounpour, Mohammad Mehdi

    2006-10-01

    Science of human identification using physiological characteristics or biometry has been of great concern in security systems. However, robust multimodal identification systems based on audio-visual information has not been thoroughly investigated yet. Therefore, the aim of this work to propose a model-based feature extraction method which employs physiological characteristics of facial muscles producing lip movements. This approach adopts the intrinsic properties of muscles such as viscosity, elasticity, and mass which are extracted from the dynamic lip model. These parameters are exclusively dependent on the neuro-muscular properties of speaker; consequently, imitation of valid speakers could be reduced to a large extent. These parameters are applied to a hidden Markov model (HMM) audio-visual identification system. In this work, a combination of audio and video features has been employed by adopting a multistream pseudo-synchronized HMM training method. Noise robust audio features such as Mel-frequency cepstral coefficients (MFCC), spectral subtraction (SS), and relative spectra perceptual linear prediction (J-RASTA-PLP) have been used to evaluate the performance of the multimodal system once efficient audio feature extraction methods have been utilized. The superior performance of the proposed system is demonstrated on a large multispeaker database of continuously spoken digits, along with a sentence that is phonetically rich. To evaluate the robustness of algorithms, some experiments were performed on genetically identical twins. Furthermore, changes in speaker voice were simulated with drug inhalation tests. In 3 dB signal to noise ratio (SNR), the dynamic muscle model improved the identification rate of the audio-visual system from 91 to 98%. Results on identical twins revealed that there was an apparent improvement on the performance for the dynamic muscle model-based system, in which the identification rate of the audio-visual system was enhanced from 87 to 96%.

  1. Education techniques for lifelong learning: giving a PowerPoint presentation: the art of communicating effectively.

    PubMed

    Collins, Jannette

    2004-01-01

    Effectiveness of an oral presentation depends on the ability of the speaker to communicate with the audience. An important part of this communication is focusing on two to five key points and emphasizing those points during the presentation. Every aspect of the presentation should be purposeful and directed at facilitating learners' achievement of the objectives. This necessitates that the speaker has carefully developed the objectives and built the presentation around attainment of the objectives. The best presentations are rehearsed, not so that the speaker memorizes exactly what he or she will say, but to facilitate the speaker's ability to interact with the audience and portray a relaxed, professional, and confident demeanor. Rehearsal also helps alleviate stage fright. The most useful method of controlling nervousness is to visualize success. When showing images, it is important to orient the audience with an adequate description, point out the relevant findings, and allow enough time for the audience to assimilate the information before moving on. This can be facilitated with appropriate use of a laser pointer, cursor, or use of builds and transitioning. A presentation should be designed to include as much audience participation as possible, no matter the size of the audience. Techniques to encourage audience participation include questioning, brainstorming, small-group activities, role-playing, case-based examples, and directed listening. It is first necessary to motivate and gain attention of the learner for learning to take place. This can be accomplished through appropriate use of humor, anecdotes, and quotations. Attention should be given to posture, body movement, eye contact, and voice when speaking, as how one appears to the audience will have an impact on their reaction to what is presented. Copyright RSNA, 2004

  2. Speech Prosody Across Stimulus Types for Individuals with Parkinson's Disease.

    PubMed

    K-Y Ma, Joan; Schneider, Christine B; Hoffmann, Rüdiger; Storch, Alexander

    2015-01-01

    Up to 89% of the individuals with Parkinson's disease (PD) experience speech problem over the course of the disease. Speech prosody and intelligibility are two of the most affected areas in hypokinetic dysarthria. However, assessment of these areas could potentially be problematic as speech prosody and intelligibility could be affected by the type of speech materials employed. To comparatively explore the effects of different types of speech stimulus on speech prosody and intelligibility in PD speakers. Speech prosody and intelligibility of two groups of individuals with varying degree of dysarthria resulting from PD was compared to that of a group of control speakers using sentence reading, passage reading and monologue. Acoustic analysis including measures on fundamental frequency (F0), intensity and speech rate was used to form a prosodic profile for each individual. Speech intelligibility was measured for the speakers with dysarthria using direct magnitude estimation. Difference in F0 variability between the speakers with dysarthria and control speakers was only observed in sentence reading task. Difference in the average intensity level was observed for speakers with mild dysarthria to that of the control speakers. Additionally, there were stimulus effect on both intelligibility and prosodic profile. The prosodic profile of PD speakers was different from that of the control speakers in the more structured task, and lower intelligibility was found in less structured task. This highlighted the value of both structured and natural stimulus to evaluate speech production in PD speakers.

  3. Effects of Language Background on Gaze Behavior: A Crosslinguistic Comparison Between Korean and German Speakers

    PubMed Central

    Goller, Florian; Lee, Donghoon; Ansorge, Ulrich; Choi, Soonja

    2017-01-01

    Languages differ in how they categorize spatial relations: While German differentiates between containment (in) and support (auf) with distinct spatial words—(a) den Kuli IN die Kappe stecken (”put pen in cap”); (b) die Kappe AUF den Kuli stecken (”put cap on pen”)—Korean uses a single spatial word (kkita) collapsing (a) and (b) into one semantic category, particularly when the spatial enclosure is tight-fit. Korean uses a different word (i.e., netha) for loose-fits (e.g., apple in bowl). We tested whether these differences influence the attention of the speaker. In a crosslinguistic study, we compared native German speakers with native Korean speakers. Participants rated the similarity of two successive video clips of several scenes where two objects were joined or nested (either in a tight or loose manner). The rating data show that Korean speakers base their rating of similarity more on tight- versus loose-fit, whereas German speakers base their rating more on containment versus support (in vs. auf). Throughout the experiment, we also measured the participants’ eye movements. Korean speakers looked equally long at the moving Figure object and at the stationary Ground object, whereas German speakers were more biased to look at the Ground object. Additionally, Korean speakers also looked more at the region where the two objects touched than did German speakers. We discuss our data in the light of crosslinguistic semantics and the extent of their influence on spatial cognition and perception. PMID:29362644

  4. Factor analysis of auto-associative neural networks with application in speaker verification.

    PubMed

    Garimella, Sri; Hermansky, Hynek

    2013-04-01

    Auto-associative neural network (AANN) is a fully connected feed-forward neural network, trained to reconstruct its input at its output through a hidden compression layer, which has fewer numbers of nodes than the dimensionality of input. AANNs are used to model speakers in speaker verification, where a speaker-specific AANN model is obtained by adapting (or retraining) the universal background model (UBM) AANN, an AANN trained on multiple held out speakers, using corresponding speaker data. When the amount of speaker data is limited, this adaptation procedure may lead to overfitting as all the parameters of UBM-AANN are adapted. In this paper, we introduce and develop the factor analysis theory of AANNs to alleviate this problem. We hypothesize that only the weight matrix connecting the last nonlinear hidden layer and the output layer is speaker-specific, and further restrict it to a common low-dimensional subspace during adaptation. The subspace is learned using large amounts of development data, and is held fixed during adaptation. Thus, only the coordinates in a subspace, also known as i-vector, need to be estimated using speaker-specific data. The update equations are derived for learning both the common low-dimensional subspace and the i-vectors corresponding to speakers in the subspace. The resultant i-vector representation is used as a feature for the probabilistic linear discriminant analysis model. The proposed system shows promising results on the NIST-08 speaker recognition evaluation (SRE), and yields a 23% relative improvement in equal error rate over the previously proposed weighted least squares-based subspace AANNs system. The experiments on NIST-10 SRE confirm that these improvements are consistent and generalize across datasets.

  5. Speaker and Accent Variation Are Handled Differently: Evidence in Native and Non-Native Listeners

    PubMed Central

    Kriengwatana, Buddhamas; Terry, Josephine; Chládková, Kateřina; Escudero, Paola

    2016-01-01

    Listeners are able to cope with between-speaker variability in speech that stems from anatomical sources (i.e. individual and sex differences in vocal tract size) and sociolinguistic sources (i.e. accents). We hypothesized that listeners adapt to these two types of variation differently because prior work indicates that adapting to speaker/sex variability may occur pre-lexically while adapting to accent variability may require learning from attention to explicit cues (i.e. feedback). In Experiment 1, we tested our hypothesis by training native Dutch listeners and Australian-English (AusE) listeners without any experience with Dutch or Flemish to discriminate between the Dutch vowels /I/ and /ε/ from a single speaker. We then tested their ability to classify /I/ and /ε/ vowels of a novel Dutch speaker (i.e. speaker or sex change only), or vowels of a novel Flemish speaker (i.e. speaker or sex change plus accent change). We found that both Dutch and AusE listeners could successfully categorize vowels if the change involved a speaker/sex change, but not if the change involved an accent change. When AusE listeners were given feedback on their categorization responses to the novel speaker in Experiment 2, they were able to successfully categorize vowels involving an accent change. These results suggest that adapting to accents may be a two-step process, whereby the first step involves adapting to speaker differences at a pre-lexical level, and the second step involves adapting to accent differences at a contextual level, where listeners have access to word meaning or are given feedback that allows them to appropriately adjust their perceptual category boundaries. PMID:27309889

  6. Integrating hidden Markov model and PRAAT: a toolbox for robust automatic speech transcription

    NASA Astrophysics Data System (ADS)

    Kabir, A.; Barker, J.; Giurgiu, M.

    2010-09-01

    An automatic time-aligned phone transcription toolbox of English speech corpora has been developed. Especially the toolbox would be very useful to generate robust automatic transcription and able to produce phone level transcription using speaker independent models as well as speaker dependent models without manual intervention. The system is based on standard Hidden Markov Models (HMM) approach and it was successfully experimented over a large audiovisual speech corpus namely GRID corpus. One of the most powerful features of the toolbox is the increased flexibility in speech processing where the speech community would be able to import the automatic transcription generated by HMM Toolkit (HTK) into a popular transcription software, PRAAT, and vice-versa. The toolbox has been evaluated through statistical analysis on GRID data which shows that automatic transcription deviates by an average of 20 ms with respect to manual transcription.

  7. Prosody in the hands of the speaker

    PubMed Central

    Guellaï, Bahia; Langus, Alan; Nespor, Marina

    2014-01-01

    In everyday life, speech is accompanied by gestures. In the present study, two experiments tested the possibility that spontaneous gestures accompanying speech carry prosodic information. Experiment 1 showed that gestures provide prosodic information, as adults are able to perceive the congruency between low-pass filtered—thus unintelligible—speech and the gestures of the speaker. Experiment 2 shows that in the case of ambiguous sentences (i.e., sentences with two alternative meanings depending on their prosody) mismatched prosody and gestures lead participants to choose more often the meaning signaled by gestures. Our results demonstrate that the prosody that characterizes speech is not a modality specific phenomenon: it is also perceived in the spontaneous gestures that accompany speech. We draw the conclusion that spontaneous gestures and speech form a single communication system where the suprasegmental aspects of spoken language are mapped to the motor-programs responsible for the production of both speech sounds and hand gestures. PMID:25071666

  8. Value of Sample Return and High Precision Analyses: Need for A Resource of Compelling Stories, Metaphors and Examples for Public Speakers

    NASA Technical Reports Server (NTRS)

    Allton, J. H.

    2017-01-01

    There is widespread agreement among planetary scientists that much of what we know about the workings of the solar system comes from accurate, high precision measurements on returned samples. Precision is a function of the number of atoms the instrumentation is able to count. Accuracy depends on the calibration or standardization technique. For Genesis, the solar wind sample return mission, acquiring enough atoms to ensure precise SW measurements and then accurately quantifying those measurements were steps known to be non-trivial pre-flight. The difficulty of precise and accurate measurements on returned samples, and why they cannot be made remotely, is not communicated well to the public. In part, this is be-cause "high precision" is abstract and error bars are not very exciting topics. This paper explores ideas for collecting and compiling compelling metaphors and colorful examples as a resource for planetary science public speakers.

  9. Malay sentiment analysis based on combined classification approaches and Senti-lexicon algorithm.

    PubMed

    Al-Saffar, Ahmed; Awang, Suryanti; Tao, Hai; Omar, Nazlia; Al-Saiagh, Wafaa; Al-Bared, Mohammed

    2018-01-01

    Sentiment analysis techniques are increasingly exploited to categorize the opinion text to one or more predefined sentiment classes for the creation and automated maintenance of review-aggregation websites. In this paper, a Malay sentiment analysis classification model is proposed to improve classification performances based on the semantic orientation and machine learning approaches. First, a total of 2,478 Malay sentiment-lexicon phrases and words are assigned with a synonym and stored with the help of more than one Malay native speaker, and the polarity is manually allotted with a score. In addition, the supervised machine learning approaches and lexicon knowledge method are combined for Malay sentiment classification with evaluating thirteen features. Finally, three individual classifiers and a combined classifier are used to evaluate the classification accuracy. In experimental results, a wide-range of comparative experiments is conducted on a Malay Reviews Corpus (MRC), and it demonstrates that the feature extraction improves the performance of Malay sentiment analysis based on the combined classification. However, the results depend on three factors, the features, the number of features and the classification approach.

  10. Malay sentiment analysis based on combined classification approaches and Senti-lexicon algorithm

    PubMed Central

    Awang, Suryanti; Tao, Hai; Omar, Nazlia; Al-Saiagh, Wafaa; Al-bared, Mohammed

    2018-01-01

    Sentiment analysis techniques are increasingly exploited to categorize the opinion text to one or more predefined sentiment classes for the creation and automated maintenance of review-aggregation websites. In this paper, a Malay sentiment analysis classification model is proposed to improve classification performances based on the semantic orientation and machine learning approaches. First, a total of 2,478 Malay sentiment-lexicon phrases and words are assigned with a synonym and stored with the help of more than one Malay native speaker, and the polarity is manually allotted with a score. In addition, the supervised machine learning approaches and lexicon knowledge method are combined for Malay sentiment classification with evaluating thirteen features. Finally, three individual classifiers and a combined classifier are used to evaluate the classification accuracy. In experimental results, a wide-range of comparative experiments is conducted on a Malay Reviews Corpus (MRC), and it demonstrates that the feature extraction improves the performance of Malay sentiment analysis based on the combined classification. However, the results depend on three factors, the features, the number of features and the classification approach. PMID:29684036

  11. Human Language Technology: Opportunities and Challenges

    DTIC Science & Technology

    2005-01-01

    because of the connections to and reliance on signal processing. Audio diarization critically includes indexing of speakers [12], since speaker ...to reduce inter- speaker variability in training. Standard techniques include vocal-tract length normalization, adaptation of acoustic models using...maximum likelihood linear regression (MLLR), and speaker -adaptive training based on MLLR. The acoustic models are mixtures of Gaussians, typically with

  12. Evaluation of Speakers with Foreign-Accented Speech in Japan: The Effect of Accent Produced by English Native Speakers

    ERIC Educational Resources Information Center

    Tsurutani, Chiharu

    2012-01-01

    Foreign-accented speakers are generally regarded as less educated, less reliable and less interesting than native speakers and tend to be associated with cultural stereotypes of their country of origin. This discrimination against foreign accents has, however, been discussed mainly using accented English in English-speaking countries. This study…

  13. The Employability of Non-Native-Speaker Teachers of EFL: A UK Survey

    ERIC Educational Resources Information Center

    Clark, Elizabeth; Paran, Amos

    2007-01-01

    The native speaker still has a privileged position in English language teaching, representing both the model speaker and the ideal teacher. Non-native-speaker teachers of English are often perceived as having a lower status than their native-speaking counterparts, and have been shown to face discriminatory attitudes when applying for teaching…

  14. Generic Language and Speaker Confidence Guide Preschoolers' Inferences about Novel Animate Kinds

    ERIC Educational Resources Information Center

    Stock, Hayli R.; Graham, Susan A.; Chambers, Craig G.

    2009-01-01

    We investigated the influence of speaker certainty on 156 four-year-old children's sensitivity to generic and nongeneric statements. An inductive inference task was implemented, in which a speaker described a nonobvious property of a novel creature using either a generic or a nongeneric statement. The speaker appeared to be confident, neutral, or…

  15. Modern Greek Language: Acquisition of Morphology and Syntax by Non-Native Speakers

    ERIC Educational Resources Information Center

    Andreou, Georgia; Karapetsas, Anargyros; Galantomos, Ioannis

    2008-01-01

    This study investigated the performance of native and non native speakers of Modern Greek language on morphology and syntax tasks. Non-native speakers of Greek whose native language was English, which is a language with strict word order and simple morphology, made more errors and answered more slowly than native speakers on morphology but not…

  16. A Comparison of Coverbal Gesture Use in Oral Discourse among Speakers with Fluent and Nonfluent Aphasia

    ERIC Educational Resources Information Center

    Kong, Anthony Pak-Hin; Law, Sam-Po; Chak, Gigi Wan-Chi

    2017-01-01

    Purpose: Coverbal gesture use, which is affected by the presence and degree of aphasia, can be culturally specific. The purpose of this study was to compare gesture use among Cantonese-speaking individuals: 23 neurologically healthy speakers, 23 speakers with fluent aphasia, and 21 speakers with nonfluent aphasia. Method: Multimedia data of…

  17. Accent Attribution in Speakers with Foreign Accent Syndrome

    ERIC Educational Resources Information Center

    Verhoeven, Jo; De Pauw, Guy; Pettinato, Michele; Hirson, Allen; Van Borsel, John; Marien, Peter

    2013-01-01

    Purpose: The main aim of this experiment was to investigate the perception of Foreign Accent Syndrome in comparison to speakers with an authentic foreign accent. Method: Three groups of listeners attributed accents to conversational speech samples of 5 FAS speakers which were embedded amongst those of 5 speakers with a real foreign accent and 5…

  18. Race in Conflict with Heritage: "Black" Heritage Language Speaker of Japanese

    ERIC Educational Resources Information Center

    Doerr, Neriko Musha; Kumagai, Yuri

    2014-01-01

    "Heritage language speaker" is a relatively new term to denote minority language speakers who grew up in a household where the language was used or those who have a family, ancestral, or racial connection to the minority language. In research on heritage language speakers, overlap between these 2 definitions is often assumed--that is,…

  19. Early Language Experience Facilitates the Processing of Gender Agreement in Spanish Heritage Speakers

    ERIC Educational Resources Information Center

    Montrul, Silvina; Davidson, Justin; De La Fuente, Israel; Foote, Rebecca

    2014-01-01

    We examined how age of acquisition in Spanish heritage speakers and L2 learners interacts with implicitness vs. explicitness of tasks in gender processing of canonical and non-canonical ending nouns. Twenty-three Spanish native speakers, 29 heritage speakers, and 33 proficiency-matched L2 learners completed three on-line spoken word recognition…

  20. The Role of Interaction in Native Speaker Comprehension of Nonnative Speaker Speech.

    ERIC Educational Resources Information Center

    Polio, Charlene; Gass, Susan M.

    1998-01-01

    Because interaction gives language learners an opportunity to modify their speech upon a signal of noncomprehension, it should also have a positive effect on native speakers' (NS) comprehension of nonnative speakers (NNS). This study shows that interaction does help NSs comprehend NNSs, contrasting the claims of an earlier study that found no…

  1. Working with Speakers.

    ERIC Educational Resources Information Center

    Pestel, Ann

    1989-01-01

    The author discusses working with speakers from business and industry to present career information at the secondary level. Advice for speakers is presented, as well as tips for program coordinators. (CH)

  2. Rhetorical structure theory and text analysis

    NASA Astrophysics Data System (ADS)

    Mann, William C.; Matthiessen, Christian M. I. M.; Thompson, Sandra A.

    1989-11-01

    Recent research on text generation has shown that there is a need for stronger linguistic theories that tell in detail how texts communicate. The prevailing theories are very difficult to compare, and it is also very difficult to see how they might be combined into stronger theories. To make comparison and combination a bit more approachable, we have created a book which is designed to encourage comparison. A dozen different authors or teams, all experienced in discourse research, are given exactly the same text to analyze. The text is an appeal for money by a lobbying organization in Washington, DC. It informs, stimulates and manipulates the reader in a fascinating way. The joint analysis is far more insightful than any one team's analysis alone. This paper is our contribution to the book. Rhetorical Structure Theory (RST), the focus of this paper, is a way to account for the functional potential of text, its capacity to achieve the purposes of speakers and produce effects in hearers. It also shows a way to distinguish coherent texts from incoherent ones, and identifies consequences of text structure.

  3. Skewed Sociolinguistic Awareness of a Native Non-standard Dialect: Evidence from the Cypriot Greek Writing of Greek Cypriot Students

    PubMed Central

    Ayiomamitou, Ioli; Yiakoumetti, Androula

    2017-01-01

    Over the last 50 years, sociolinguistic research in settings in which a regional, social, or ethnic non-standard linguistic variety is used alongside the standard variety of the same language has steadily increased. The educational implications of the concomitant use of such varieties have also received a great deal of research attention. This study deals with regional linguistic variation and its implications for education by focusing on the Greek Cypriot educational context. This context is ideal for investigating the linguistic profiles of speakers of proximal varieties as the majority of Greek Cypriots are primarily educated in just one of their varieties: the standard educational variety. The aim of our study was to understand Greek Cypriot primary school pupils’ sociolinguistic awareness via examination of their written production in their home variety [Cypriot Greek (CG) dialect]. Our assumption was that, because written production is less spontaneous than speech, it better reflects pupils’ conscious awareness. Pupils were advised to produce texts that reflected their everyday language with family and friends (beyond school boundaries). As expected, students’ texts included an abundance of mesolectal features and the following were the ten most frequent: (1) palato-alveolar consonants, (2) future particle [ená] and conditional [ítan na] + subjunctive, (3) consonant devoicing, (4) CG-specific verb stems, (5) final [n] retention, (6) [én/ éni] instead of [íne], (7) CG-specific verb endings, (8) [én/é] instead of [ðen], (9) elision of intervocalic fricative [ɣ], and (10) CG-specific adverbs. Importantly, in addition to the expected mesolectal features that reflect contemporary CG, students included a significant and unexpected number of basilectal features and instances of hyperdialectism (that are not representative of today’s linguistic reality) which rendered their texts register-inappropriate. This led us to conclude that Greek Cypriot students have a skewed sociolinguistic awareness of variation within their first dialect and a distorted impression of their own everyday language. We argue that the portrayal of CG in its basilectal form was performed intentionally by students in an effort to distance themselves from a socially constructed identity of a rural, uneducated, and stigmatized non-standard-dialect speaker. The study is of international relevance as it deals with sociolinguistic issues that pertain to all bidialectal speakers. PMID:29213249

  4. Variability and Intelligibility of Clarified Speech to Different Listener Groups

    NASA Astrophysics Data System (ADS)

    Silber, Ronnie F.

    Two studies examined the modifications that adult speakers make in speech to disadvantaged listeners. Previous research that has focused on speech to the deaf individuals and to young children has shown that adults clarify speech when addressing these two populations. Acoustic measurements suggest that the signal undergoes similar changes for both populations. Perceptual tests corroborate these results for the deaf population, but are nonsystematic in developmental studies. The differences in the findings for these populations and the nonsystematic results in the developmental literature may be due to methodological factors. The present experiments addressed these methodological questions. Studies of speech to hearing impaired listeners have used read, nonsense, sentences, for which speakers received explicit clarification instructions and feedback, while in the child literature, excerpts of real-time conversations were used. Therefore, linguistic samples were not precisely matched. In this study, experiments used various linguistic materials. Experiment 1 used a children's story; experiment 2, nonsense sentences. Four mothers read both types of material in four ways: (1) in "normal" adult speech, (2) in "babytalk," (3) under the clarification instructions used in the "hearing impaired studies" (instructed clear speech) and (4) in (spontaneous) clear speech without instruction. No extra practice or feedback was given. Sentences were presented to 40 normal hearing college students with and without simultaneous masking noise. Results were separately tabulated for content and function words, and analyzed using standard statistical tests. The major finding in the study was individual variation in speaker intelligibility. "Real world" speakers vary in their baseline intelligibility. The four speakers also showed unique patterns of intelligibility as a function of each independent variable. Results were as follows. Nonsense sentences were less intelligible than story sentences. Function words were equal to, or more intelligible than, content words. Babytalk functioned as a clear speech style in story sentences but not nonsense sentences. One of the two clear speech styles was clearer than normal speech in adult-directed clarification. However, which style was clearer depended on interactions among the variables. The individual patterns seemed to result from interactions among demand characteristics, baseline intelligibility, materials, and differences in articulatory flexibility.

  5. Grammatical Planning Units During Real-Time Sentence Production in Speakers With Agrammatic Aphasia and Healthy Speakers.

    PubMed

    Lee, Jiyeon; Yoshida, Masaya; Thompson, Cynthia K

    2015-08-01

    Grammatical encoding (GE) is impaired in agrammatic aphasia; however, the nature of such deficits remains unclear. We examined grammatical planning units during real-time sentence production in speakers with agrammatic aphasia and control speakers, testing two competing models of GE. We queried whether speakers with agrammatic aphasia produce sentences word by word without advanced planning or whether hierarchical syntactic structure (i.e., verb argument structure; VAS) is encoded as part of the advanced planning unit. Experiment 1 examined production of sentences with a predefined structure (i.e., "The A and the B are above the C") using eye tracking. Experiment 2 tested production of transitive and unaccusative sentences without a predefined sentence structure in a verb-priming study. In Experiment 1, both speakers with agrammatic aphasia and young and age-matched control speakers used word-by-word strategies, selecting the first lemma (noun A) only prior to speech onset. However, in Experiment 2, unlike controls, speakers with agrammatic aphasia preplanned transitive and unaccusative sentences, encoding VAS before speech onset. Speakers with agrammatic aphasia show incremental, word-by-word production for structurally simple sentences, requiring retrieval of multiple noun lemmas. However, when sentences involve functional (thematic to grammatical) structure building, advanced planning strategies (i.e., VAS encoding) are used. This early use of hierarchical syntactic information may provide a scaffold for impaired GE in agrammatism.

  6. Grammatical Encoding and Learning in Agrammatic Aphasia: Evidence from Structural Priming

    PubMed Central

    Cho-Reyes, Soojin; Mack, Jennifer E.; Thompson, Cynthia K.

    2017-01-01

    The present study addressed open questions about the nature of sentence production deficits in agrammatic aphasia. In two structural priming experiments, 13 aphasic and 13 age-matched control speakers repeated visually- and auditorily-presented prime sentences, and then used visually-presented word arrays to produce dative sentences. Experiment 1 examined whether agrammatic speakers form structural and thematic representations during sentence production, whereas Experiment 2 tested the lasting effects of structural priming in lags of two and four sentences. Results of Experiment 1 showed that, like unimpaired speakers, the aphasic speakers evinced intact structural priming effects, suggesting that they are able to generate such representations. Unimpaired speakers also evinced reliable thematic priming effects, whereas agrammatic speakers did so in some experimental conditions, suggesting that access to thematic representations may be intact. Results of Experiment 2 showed structural priming effects of comparable magnitude for aphasic and unimpaired speakers. In addition, both groups showed lasting structural priming effects in both lag conditions, consistent with implicit learning accounts. In both experiments, aphasic speakers with more severe language impairments exhibited larger priming effects, consistent with the “inverse preference” prediction of implicit learning accounts. The findings indicate that agrammatic speakers are sensitive to structural priming across levels of representation and that such effects are lasting, suggesting that structural priming may be beneficial for the treatment of sentence production deficits in agrammatism. PMID:28924328

  7. Brief Report: Relations between Prosodic Performance and Communication and Socialization Ratings in High Functioning Speakers with Autism Spectrum Disorders

    ERIC Educational Resources Information Center

    Paul, Rhea; Shriberg, Lawrence D.; McSweeny, Jane; Cicchetti, Domenic; Klin, Ami; Volkmar, Fred

    2005-01-01

    Shriberg "et al." [Shriberg, L. "et al." (2001). "Journal of Speech, Language and Hearing Research, 44," 1097-1115] described prosody-voice features of 30 high functioning speakers with autistic spectrum disorder (ASD) compared to age-matched control speakers. The present study reports additional information on the speakers with ASD, including…

  8. Investigating Holistic Measures of Speech Prosody

    ERIC Educational Resources Information Center

    Cunningham, Dana Aliel

    2012-01-01

    Speech prosody is a multi-faceted dimension of speech which can be measured and analyzed in a variety of ways. In this study, the speech prosody of Mandarin L1 speakers, English L2 speakers, and English L1 speakers was assessed by trained raters who listened to sound clips of the speakers responding to a graph prompt and reading a short passage.…

  9. Young Children's Sensitivity to Speaker Gender When Learning from Others

    ERIC Educational Resources Information Center

    Ma, Lili; Woolley, Jacqueline D.

    2013-01-01

    This research explores whether young children are sensitive to speaker gender when learning novel information from others. Four- and 6-year-olds ("N" = 144) chose between conflicting statements from a male versus a female speaker (Studies 1 and 3) or decided which speaker (male or female) they would ask (Study 2) when learning about the functions…

  10. Switches to English during French Service Encounters: Relationships with L2 French Speakers' Willingness to Communicate and Motivation

    ERIC Educational Resources Information Center

    McNaughton, Stephanie; McDonough, Kim

    2015-01-01

    This exploratory study investigated second language (L2) French speakers' service encounters in the multilingual setting of Montreal, specifically whether switches to English during French service encounters were related to L2 speakers' willingness to communicate or motivation. Over a two-week period, 17 French L2 speakers in Montreal submitted…

  11. A Respirometric Technique to Evaluate Velopharyngeal Function in Speakers with Cleft Palate, with and without Prostheses.

    ERIC Educational Resources Information Center

    Gilbert, Harvey R.; Ferrand, Carole T.

    1987-01-01

    Respirometric quotients (RQ), the ratio of oral air volume expended to total volume expended, were obtained from the productions of oral and nasal airflow of 10 speakers with cleft palate, with and without their prosthetic appliances, and 10 normal speakers. Cleft palate speakers without their appliances exhibited the lowest RQ values. (Author/DB)

  12. Using Stimulated Recall to Investigate Native Speaker Perceptions in Native-Nonnative Speaker Interaction

    ERIC Educational Resources Information Center

    Polio, Charlene; Gass, Susan; Chapin, Laura

    2006-01-01

    Implicit negative feedback has been shown to facilitate SLA, and the extent to which such feedback is given is related to a variety of task and interlocutor variables. The background of a native speaker (NS), in terms of amount of experience in interactions with nonnative speakers (NNSs), has been shown to affect the quantity of implicit negative…

  13. Compliment Responses: Comparing American Learners of Japanese, Native Japanese Speakers, and American Native English Speakers

    ERIC Educational Resources Information Center

    Tatsumi, Naofumi

    2012-01-01

    Previous research shows that American learners of Japanese (AJs) tend to differ from native Japanese speakers in their compliment responses (CRs). Yokota (1986) and Shimizu (2009) have reported that AJs tend to respond more negatively than native Japanese speakers. It has also been reported that AJs' CRs tend to lack the use of avoidance or…

  14. Factors Influencing Oral Corrective Feedback Provision in the Spanish Foreign Language Classroom: Investigating Instructor Native/Nonnative Speaker Status, SLA Education, & Teaching Experience

    ERIC Educational Resources Information Center

    Gurzynski-Weiss, Laura

    2010-01-01

    The role of interactional feedback has been a critical area of second language acquisition (SLA) research for decades and while findings suggest interactional feedback can facilitate SLA, the extent of its influence can vary depending on a number of factors, including the native language of those involved in communication. Although studies have…

  15. Intonation contrast in Cantonese speakers with hypokinetic dysarthria associated with Parkinson's disease.

    PubMed

    Ma, Joan K-Y; Whitehill, Tara L; So, Susanne Y-S

    2010-08-01

    Speech produced by individuals with hypokinetic dysarthria associated with Parkinson's disease (PD) is characterized by a number of features including impaired speech prosody. The purpose of this study was to investigate intonation contrasts produced by this group of speakers. Speech materials with a question-statement contrast were collected from 14 Cantonese speakers with PD. Twenty listeners then classified the productions as either questions or statements. Acoustic analyses of F0, duration, and intensity were conducted to determine which acoustic cues distinguished the production of questions from statements, and which cues appeared to be exploited by listeners in identifying intonational contrasts. The results show that listeners identified statements with a high degree of accuracy, but the accuracy of question identification ranged from 0.56% to 96% across the 14 speakers. The speakers with PD used similar acoustic cues as nondysarthric Cantonese speakers to mark the question-statement contrast, although the contrasts were not observed in all speakers. Listeners mainly used F0 cues at the final syllable for intonation identification. These data contribute to the researchers' understanding of intonation marking in speakers with PD, with specific application to the production and perception of intonation in a lexical tone language.

  16. Intelligibility of clear speech: effect of instruction.

    PubMed

    Lam, Jennifer; Tjaden, Kris

    2013-10-01

    The authors investigated how clear speech instructions influence sentence intelligibility. Twelve speakers produced sentences in habitual, clear, hearing impaired, and overenunciate conditions. Stimuli were amplitude normalized and mixed with multitalker babble for orthographic transcription by 40 listeners. The main analysis investigated percentage-correct intelligibility scores as a function of the 4 conditions and speaker sex. Additional analyses included listener response variability, individual speaker trends, and an alternate intelligibility measure: proportion of content words correct. Relative to the habitual condition, the overenunciate condition was associated with the greatest intelligibility benefit, followed by the hearing impaired and clear conditions. Ten speakers followed this trend. The results indicated different patterns of clear speech benefit for male and female speakers. Greater listener variability was observed for speakers with inherently low habitual intelligibility compared to speakers with inherently high habitual intelligibility. Stable proportions of content words were observed across conditions. Clear speech instructions affected the magnitude of the intelligibility benefit. The instruction to overenunciate may be most effective in clear speech training programs. The findings may help explain the range of clear speech intelligibility benefit previously reported. Listener variability analyses suggested the importance of obtaining multiple listener judgments of intelligibility, especially for speakers with inherently low habitual intelligibility.

  17. Discrimination of speaker sex and size when glottal-pulse rate and vocal-tract length are controlled.

    PubMed

    Smith, David R R; Walters, Thomas C; Patterson, Roy D

    2007-12-01

    A recent study [Smith and Patterson, J. Acoust. Soc. Am. 118, 3177-3186 (2005)] demonstrated that both the glottal-pulse rate (GPR) and the vocal-tract length (VTL) of vowel sounds have a large effect on the perceived sex and age (or size) of a speaker. The vowels for all of the "different" speakers in that study were synthesized from recordings of the sustained vowels of one, adult male speaker. This paper presents a follow-up study in which a range of vowels were synthesized from recordings of four different speakers--an adult man, an adult woman, a young boy, and a young girl--to determine whether the sex and age of the original speaker would have an effect upon listeners' judgments of whether a vowel was spoken by a man, woman, boy, or girl, after they were equated for GPR and VTL. The sustained vowels of the four speakers were scaled to produce the same combinations of GPR and VTL, which covered the entire range normally encountered in every day life. The results show that listeners readily distinguish children from adults based on their sustained vowels but that they struggle to distinguish the sex of the speaker.

  18. On the same wavelength: predictable language enhances speaker-listener brain-to-brain synchrony in posterior superior temporal gyrus.

    PubMed

    Dikker, Suzanne; Silbert, Lauren J; Hasson, Uri; Zevin, Jason D

    2014-04-30

    Recent research has shown that the degree to which speakers and listeners exhibit similar brain activity patterns during human linguistic interaction is correlated with communicative success. Here, we used an intersubject correlation approach in fMRI to test the hypothesis that a listener's ability to predict a speaker's utterance increases such neural coupling between speakers and listeners. Nine subjects listened to recordings of a speaker describing visual scenes that varied in the degree to which they permitted specific linguistic predictions. In line with our hypothesis, the temporal profile of listeners' brain activity was significantly more synchronous with the speaker's brain activity for highly predictive contexts in left posterior superior temporal gyrus (pSTG), an area previously associated with predictive auditory language processing. In this region, predictability differentially affected the temporal profiles of brain responses in the speaker and listeners respectively, in turn affecting correlated activity between the two: whereas pSTG activation increased with predictability in the speaker, listeners' pSTG activity instead decreased for more predictable sentences. Listeners additionally showed stronger BOLD responses for predictive images before sentence onset, suggesting that highly predictable contexts lead comprehenders to preactivate predicted words.

  19. When speaker identity is unavoidable: Neural processing of speaker identity cues in natural speech.

    PubMed

    Tuninetti, Alba; Chládková, Kateřina; Peter, Varghese; Schiller, Niels O; Escudero, Paola

    2017-11-01

    Speech sound acoustic properties vary largely across speakers and accents. When perceiving speech, adult listeners normally disregard non-linguistic variation caused by speaker or accent differences, in order to comprehend the linguistic message, e.g. to correctly identify a speech sound or a word. Here we tested whether the process of normalizing speaker and accent differences, facilitating the recognition of linguistic information, is found at the level of neural processing, and whether it is modulated by the listeners' native language. In a multi-deviant oddball paradigm, native and nonnative speakers of Dutch were exposed to naturally-produced Dutch vowels varying in speaker, sex, accent, and phoneme identity. Unexpectedly, the analysis of mismatch negativity (MMN) amplitudes elicited by each type of change shows a large degree of early perceptual sensitivity to non-linguistic cues. This finding on perception of naturally-produced stimuli contrasts with previous studies examining the perception of synthetic stimuli wherein adult listeners automatically disregard acoustic cues to speaker identity. The present finding bears relevance to speech normalization theories, suggesting that at an unattended level of processing, listeners are indeed sensitive to changes in fundamental frequency in natural speech tokens. Copyright © 2017 Elsevier Inc. All rights reserved.

  20. Patterns of lung volume use during an extemporaneous speech task in persons with Parkinson disease.

    PubMed

    Bunton, Kate

    2005-01-01

    This study examined patterns of lung volume use in speakers with Parkinson disease (PD) during an extemporaneous speaking task. The performance of a control group was also examined. Behaviors described are based on acoustic, kinematic and linguistic measures. Group differences were found in breath group duration, lung volume initiation, and lung volume termination measures. Speakers in the control group alternated between a longer and shorter breath groups. With starting lung volumes being higher for the longer breath groups and lower for shorter breath groups. Speech production was terminated before reaching tidal end expiratory level. This pattern was also seen in 4 of 7 speakers with PD. The remaining 3 PD speakers initiated speech at low starting lung volumes and continued speaking below EEL. This subgroup of PD speakers ended breath groups at agrammatical boundaries, whereas control speakers ended at appropriate grammatical boundaries. As a result of participating in this exercise, the reader will (1) be able to describe the patterns of lung volume use in speakers with Parkinson disease and compare them with those employed by control speakers; and (2) obtain information about the influence of speaking task on speech breathing.

  1. When one person's mistake is another's standard usage: the effect of foreign accent on syntactic processing.

    PubMed

    Hanulíková, Adriana; van Alphen, Petra M; van Goch, Merel M; Weber, Andrea

    2012-04-01

    How do native listeners process grammatical errors that are frequent in non-native speech? We investigated whether the neural correlates of syntactic processing are modulated by speaker identity. ERPs to gender agreement errors in sentences spoken by a native speaker were compared with the same errors spoken by a non-native speaker. In line with previous research, gender violations in native speech resulted in a P600 effect (larger P600 for violations in comparison with correct sentences), but when the same violations were produced by the non-native speaker with a foreign accent, no P600 effect was observed. Control sentences with semantic violations elicited comparable N400 effects for both the native and the non-native speaker, confirming no general integration problem in foreign-accented speech. The results demonstrate that the P600 is modulated by speaker identity, extending our knowledge about the role of speaker's characteristics on neural correlates of speech processing.

  2. Factors affecting the perception of Korean-accented American English

    NASA Astrophysics Data System (ADS)

    Cho, Kwansun; Harris, John G.; Shrivastav, Rahul

    2005-09-01

    This experiment examines the relative contribution of two factors, intonation and articulation errors, on the perception of foreign accent in Korean-accented American English. Ten native speakers of Korean and ten native speakers of American English were asked to read ten English sentences. These sentences were then modified using high-quality speech resynthesis techniques [STRAIGHT Kawahara et al., Speech Commun. 27, 187-207 (1999)] to generate four sets of stimuli. In the first two sets of stimuli, the intonation patterns of the Korean speakers and American speakers were switched with one another. The articulatory errors for each speaker were not modified. In the final two sets, the sentences from the Korean and American speakers were resynthesized without any modifications. Fifteen listeners were asked to rate all the stimuli for the degree of foreign accent. Preliminary results show that, for native speakers of American English, articulation errors may play a greater role in the perception of foreign accent than errors in intonation patterns. [Work supported by KAIM.

  3. The Role of Oral Language Skills in Reading and Listening Comprehension of Text: A Comparison of Monolingual (L1) and Bilingual (L2) Speakers of English Language

    ERIC Educational Resources Information Center

    Babayigit, Selma

    2014-01-01

    The study examined the role of oral language skills in reading comprehension and listening comprehension levels of 125 monolingual (L1) and bilingual (L2) English-speaking learners (M = 121.5 months, SD = 4.65) in England. All testing was conducted in English. The L1 learners outperformed their L2 peers on the measures of oral language and text…

  4. RTP Speakers Bureau

    EPA Pesticide Factsheets

    The Research Triangle Park Speakers Bureau page is a free resource that schools, universities, and community groups in the Raleigh-Durham-Chapel Hill, N.C. area can use to request speakers and find educational resources.

  5. Children's Understanding That Utterances Emanate from Minds: Using Speaker Belief To Aid Interpretation.

    ERIC Educational Resources Information Center

    Mitchell, Peter; Robinson, Elizabeth J.; Thompson, Doreen E.

    1999-01-01

    Three experiments examined 3- to 6-year olds' ability to use a speaker's utterance based on false belief to identify which of several referents was intended. Found that many 4- to 5-year olds performed correctly only when it was unnecessary to consider the speaker's belief. When the speaker gave an ambiguous utterance, many 3- to 6-year olds…

  6. Speaker Introductions at Internal Medicine Grand Rounds: Forms of Address Reveal Gender Bias.

    PubMed

    Files, Julia A; Mayer, Anita P; Ko, Marcia G; Friedrich, Patricia; Jenkins, Marjorie; Bryan, Michael J; Vegunta, Suneela; Wittich, Christopher M; Lyle, Melissa A; Melikian, Ryan; Duston, Trevor; Chang, Yu-Hui H; Hayes, Sharonne N

    2017-05-01

    Gender bias has been identified as one of the drivers of gender disparity in academic medicine. Bias may be reinforced by gender subordinating language or differential use of formality in forms of address. Professional titles may influence the perceived expertise and authority of the referenced individual. The objective of this study is to examine how professional titles were used in the same and mixed-gender speaker introductions at Internal Medicine Grand Rounds (IMGR). A retrospective observational study of video-archived speaker introductions at consecutive IMGR was conducted at two different locations (Arizona, Minnesota) of an academic medical center. Introducers and speakers at IMGR were physician and scientist peers holding MD, PhD, or MD/PhD degrees. The primary outcome was whether or not a speaker's professional title was used during the first form of address during speaker introductions at IMGR. As secondary outcomes, we evaluated whether or not the speakers professional title was used in any form of address during the introduction. Three hundred twenty-one forms of address were analyzed. Female introducers were more likely to use professional titles when introducing any speaker during the first form of address compared with male introducers (96.2% [102/106] vs. 65.6% [141/215]; p < 0.001). Female dyads utilized formal titles during the first form of address 97.8% (45/46) compared with male dyads who utilized a formal title 72.4% (110/152) of the time (p = 0.007). In mixed-gender dyads, where the introducer was female and speaker male, formal titles were used 95.0% (57/60) of the time. Male introducers of female speakers utilized professional titles 49.2% (31/63) of the time (p < 0.001). In this study, women introduced by men at IMGR were less likely to be addressed by professional title than were men introduced by men. Differential formality in speaker introductions may amplify isolation, marginalization, and professional discomfiture expressed by women faculty in academic medicine.

  7. Evaluating the lexico-grammatical differences in the writing of native and non-native speakers of English in peer-reviewed medical journals in the field of pediatric oncology: Creation of the genuine index scoring system.

    PubMed

    Gayle, Alberto Alexander; Shimaoka, Motomu

    2017-01-01

    The predominance of English in scientific research has created hurdles for "non-native speakers" of English. Here we present a novel application of native language identification (NLI) for the assessment of medical-scientific writing. For this purpose, we created a novel classification system whereby scoring would be based solely on text features found to be distinctive among native English speakers (NS) within a given context. We dubbed this the "Genuine Index" (GI). This methodology was validated using a small set of journals in the field of pediatric oncology. Our dataset consisted of 5,907 abstracts, representing work from 77 countries. A support vector machine (SVM) was used to generate our model and for scoring. Accuracy, precision, and recall of the classification model were 93.3%, 93.7%, and 99.4%, respectively. Class specific F-scores were 96.5% for NS and 39.8% for our benchmark class, Japan. Overall kappa was calculated to be 37.2%. We found significant differences between countries with respect to the GI score. Significant correlation was found between GI scores and two validated objective measures of writing proficiency and readability. Two sets of key terms and phrases differentiating NS and non-native writing were identified. Our GI model was able to detect, with a high degree of reliability, subtle differences between the terms and phrasing used by native and non-native speakers in peer reviewed journals, in the field of pediatric oncology. In addition, L1 language transfer was found to be very likely to survive revision, especially in non-Western countries such as Japan. These findings show that even when the language used is technically correct, there may still be some phrasing or usage that impact quality.

  8. Infants' understanding of false labeling events: the referential roles of words and the speakers who use them.

    PubMed

    Koenig, Melissa A; Echols, Catharine H

    2003-04-01

    The four studies reported here examine whether 16-month-old infants' responses to true and false utterances interact with their knowledge of human agents. In Study 1, infants heard repeated instances either of true or false labeling of common objects; labels came from an active human speaker seated next to the infant. In Study 2, infants experienced the same stimuli and procedure; however, we replaced the human speaker of Study 1 with an audio speaker in the same location. In Study 3, labels came from a hidden audio speaker. In Study 4, a human speaker labeled the objects while facing away from them. In Study 1, infants looked significantly longer to the human agent when she falsely labeled than when she truthfully labeled the objects. Infants did not show a similar pattern of attention for the audio speaker of Study 2, the silent human of Study 3 or the facing-backward speaker of Study 4. In fact, infants who experienced truthful labeling looked significantly longer to the facing-backward labeler of Study 4 than to true labelers of the other three contexts. Additionally, infants were more likely to correct false labels when produced by the human labeler of Study 1 than in any of the other contexts. These findings suggest, first, that infants are developing a critical conception of other human speakers as truthful communicators, and second, that infants understand that human speakers may provide uniquely useful information when a word fails to match its referent. These findings are consistent with the view that infants can recognize differences in knowledge and that such differences can be based on differences in the availability of perceptual experience.

  9. Request a Speaker

    Science.gov Websites

    . Northern Command Speakers Program The U.S. Northern Command Speaker's Program works to increase face-to -face contact with our public to help build and sustain public understanding of our command missions and

  10. Speakers of Different Languages Process the Visual World Differently

    PubMed Central

    Chabal, Sarah; Marian, Viorica

    2015-01-01

    Language and vision are highly interactive. Here we show that people activate language when they perceive the visual world, and that this language information impacts how speakers of different languages focus their attention. For example, when searching for an item (e.g., clock) in the same visual display, English and Spanish speakers look at different objects. Whereas English speakers searching for the clock also look at a cloud, Spanish speakers searching for the clock also look at a gift, because the Spanish names for gift (regalo) and clock (reloj) overlap phonologically. These different looking patterns emerge despite an absence of direct linguistic input, showing that language is automatically activated by visual scene processing. We conclude that the varying linguistic information available to speakers of different languages affects visual perception, leading to differences in how the visual world is processed. PMID:26030171

  11. Learning foreign labels from a foreign speaker: the role of (limited) exposure to a second language.

    PubMed

    Akhtar, Nameera; Menjivar, Jennifer; Hoicka, Elena; Sabbagh, Mark A

    2012-11-01

    Three- and four-year-olds (N = 144) were introduced to novel labels by an English speaker and a foreign speaker (of Nordish, a made-up language), and were asked to endorse one of the speaker's labels. Monolingual English-speaking children were compared to bilingual children and English-speaking children who were regularly exposed to a language other than English. All children tended to endorse the English speaker's labels when asked 'What do you call this?', but when asked 'What do you call this in Nordish?', children with exposure to a second language were more likely to endorse the foreign label than monolingual and bilingual children. The findings suggest that, at this age, exposure to, but not necessarily immersion in, more than one language may promote the ability to learn foreign words from a foreign speaker.

  12. Surmounting the Tower of Babel: Monolingual and bilingual 2-year-olds' understanding of the nature of foreign language words.

    PubMed

    Byers-Heinlein, Krista; Chen, Ke Heng; Xu, Fei

    2014-03-01

    Languages function as independent and distinct conventional systems, and so each language uses different words to label the same objects. This study investigated whether 2-year-old children recognize that speakers of their native language and speakers of a foreign language do not share the same knowledge. Two groups of children unfamiliar with Mandarin were tested: monolingual English-learning children (n=24) and bilingual children learning English and another language (n=24). An English speaker taught children the novel label fep. On English mutual exclusivity trials, the speaker asked for the referent of a novel label (wug) in the presence of the fep and a novel object. Both monolingual and bilingual children disambiguated the reference of the novel word using a mutual exclusivity strategy, choosing the novel object rather than the fep. On similar trials with a Mandarin speaker, children were asked to find the referent of a novel Mandarin label kuò. Monolinguals again chose the novel object rather than the object with the English label fep, even though the Mandarin speaker had no access to conventional English words. Bilinguals did not respond systematically to the Mandarin speaker, suggesting that they had enhanced understanding of the Mandarin speaker's ignorance of English words. The results indicate that monolingual children initially expect words to be conventionally shared across all speakers-native and foreign. Early bilingual experience facilitates children's discovery of the nature of foreign language words. Copyright © 2013 Elsevier Inc. All rights reserved.

  13. Content-specific coordination of listeners' to speakers' EEG during communication.

    PubMed

    Kuhlen, Anna K; Allefeld, Carsten; Haynes, John-Dylan

    2012-01-01

    Cognitive neuroscience has recently begun to extend its focus from the isolated individual mind to two or more individuals coordinating with each other. In this study we uncover a coordination of neural activity between the ongoing electroencephalogram (EEG) of two people-a person speaking and a person listening. The EEG of one set of twelve participants ("speakers") was recorded while they were narrating short stories. The EEG of another set of twelve participants ("listeners") was recorded while watching audiovisual recordings of these stories. Specifically, listeners watched the superimposed videos of two speakers simultaneously and were instructed to attend either to one or the other speaker. This allowed us to isolate neural coordination due to processing the communicated content from the effects of sensory input. We find several neural signatures of communication: First, the EEG is more similar among listeners attending to the same speaker than among listeners attending to different speakers, indicating that listeners' EEG reflects content-specific information. Secondly, listeners' EEG activity correlates with the attended speakers' EEG, peaking at a time delay of about 12.5 s. This correlation takes place not only between homologous, but also between non-homologous brain areas in speakers and listeners. A semantic analysis of the stories suggests that listeners coordinate with speakers at the level of complex semantic representations, so-called "situation models". With this study we link a coordination of neural activity between individuals directly to verbally communicated information.

  14. The ICSI+ Multilingual Sentence Segmentation System

    DTIC Science & Technology

    2006-01-01

    these steps the ASR output needs to be enriched with information additional to words, such as speaker diarization , sentence segmentation, or story...and the out- of a speaker diarization is considered as well. We first detail extraction of the prosodic features, and then describe the clas- ation...also takes into account the speaker turns that estimated by the diarization system. In addition to the Max- 1) model speaker turn unigrams, trigram

  15. Speaker Segmentation and Clustering Using Gender Information

    DTIC Science & Technology

    2006-02-01

    used in the first stages of segmentation forder information in the clustering of the opposite-gender speaker diarization of news broadcasts. files, the...AFRL-HE-WP-TP-2006-0026 AIR FORCE RESEARCH LABORATORY Speaker Segmentation and Clustering Using Gender Information Brian M. Ore General Dynamics...COVERED (From - To) February 2006 ProceedinLgs 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER Speaker Segmentation and Clustering Using Gender Information 5b

  16. Magnetic Fluids Deliver Better Speaker Sound Quality

    NASA Technical Reports Server (NTRS)

    2015-01-01

    In the 1960s, Glenn Research Center developed a magnetized fluid to draw rocket fuel into spacecraft engines while in space. Sony has incorporated the technology into its line of slim speakers by using the fluid as a liquid stand-in for the speaker's dampers, which prevent the speaker from blowing out while adding stability. The fluid helps to deliver more volume and hi-fidelity sound while reducing distortion.

  17. Special Observance Planning Guide

    DTIC Science & Technology

    2015-11-01

    Finding the right speaker for an event can be a challenge. Many speakers are recommended based on word-of-mouth or through a group connected to...An unprepared, rambling speaker or one who intentionally or unintentionally attacks a group or its members can be extremely damaging to a program...Don’t assume that an organizational senior leader is an adequate speaker based on position, rank, and/or affiliation with a reference group

  18. Coronal View Ultrasound Imaging of Movement in Different Segments of the Tongue during Paced Recital: Findings from Four Normal Speakers and a Speaker with Partial Glossectomy

    ERIC Educational Resources Information Center

    Bressmann, Tim; Flowers, Heather; Wong, Willy; Irish, Jonathan C.

    2010-01-01

    The goal of this study was to quantitatively describe aspects of coronal tongue movement in different anatomical regions of the tongue. Four normal speakers and a speaker with partial glossectomy read four repetitions of a metronome-paced poem. Their tongue movement was recorded in four coronal planes using two-dimensional B-mode ultrasound…

  19. Making Math Real: Effective Qualities of Guest Speaker Presentations and the Impact of Speakers on Student Attitude and Achievement in the Algebra Classroom

    ERIC Educational Resources Information Center

    McKain, Danielle R.

    2012-01-01

    The term real world is often used in mathematics education, yet the definition of real-world problems and how to incorporate them in the classroom remains ambiguous. One way real-world connections can be made is through guest speakers. Guest speakers can offer different perspectives and share knowledge about various subject areas, yet the impact…

  20. When pitch Accents Encode Speaker Commitment: Evidence from French Intonation.

    PubMed

    Michelas, Amandine; Portes, Cristel; Champagne-Lavau, Maud

    2016-06-01

    Recent studies on a variety of languages have shown that a speaker's commitment to the propositional content of his or her utterance can be encoded, among other strategies, by pitch accent types. Since prior research mainly relied on lexical-stress languages, our understanding of how speakers of a non-lexical-stress language encode speaker commitment is limited. This paper explores the contribution of the last pitch accent of an intonation phrase to convey speaker commitment in French, a language that has stress at the phrasal level as well as a restricted set of pitch accents. In a production experiment, participants had to produce sentences in two pragmatic contexts: unbiased questions (the speaker had no particular belief with respect to the expected answer) and negatively biased questions (the speaker believed the proposition to be false). Results revealed that negatively biased questions consistently exhibited an additional unaccented F0 peak in the preaccentual syllable (an H+!H* pitch accent) while unbiased questions were often realized with a rising pattern across the accented syllable (an H* pitch accent). These results provide evidence that pitch accent types in French can signal the speaker's belief about the certainty of the proposition expressed in French. It also has implications for the phonological model of French intonation.

  1. Sociological effects on vocal aging: Age related F0 effects in two languages

    NASA Astrophysics Data System (ADS)

    Nagao, Kyoko

    2005-04-01

    Listeners can estimate the age of a speaker fairly accurately from their speech (Ptacek and Sander, 1966). It is generally considered that this perception is based on physiologically determined aspects of the speech. However, the degree to which it is due to conventional sociolinguistic aspects of speech is unknown. The current study examines the degree to which fundamental frequency (F0) changes due to advanced aging across two language groups of speakers. It also examines the degree to which the speakers associate these changes with aging in a voice disguising task. Thirty native speakers each of English and Japanese, taken from three age groups, read a target phrase embedded in a carrier sentence in their native language. Each speaker also read the sentence pretending to be 20-years younger or 20-years older than their own age. Preliminary analysis of eighteen Japanese speakers indicates that the mean and maximum F0 values increase when the speakers pretended to be younger than when they pretended to be older. Some previous studies on age perception, however, suggested that F0 has minor effects on listeners' age estimation. The acoustic results will also be discussed in conjunction with the results of the listeners' age estimation of the speakers.

  2. Challenging stereotypes and changing attitudes: Improving quality of care for people with hepatitis C through Positive Speakers programs.

    PubMed

    Brener, Loren; Wilson, Hannah; Rose, Grenville; Mackenzie, Althea; de Wit, John

    2013-01-01

    Positive Speakers programs consist of people who are trained to speak publicly about their illness. The focus of these programs, especially with stigmatised illnesses such as hepatitis C (HCV), is to inform others of the speakers' experiences, thereby humanising the illness and reducing ignorance associated with the disease. This qualitative research aimed to understand the perceived impact of Positive Speakers programs on changing audience members' attitudes towards people with HCV. Interviews were conducted with nine Positive Speakers and 16 of their audience members to assess the way in which these sessions were perceived by both speakers and the audience to challenge stereotypes and stigma associated with HCV and promote positive attitude change amongst the audience. Data were analysed using Intergroup Contact Theory to frame the analysis with a focus on whether the program met the optimal conditions to promote attitude change. Findings suggest that there are a number of vital components to this Positive Speakers program which ensures that the program meets the requirements for successful and equitable intergroup contact. This Positive Speakers program thereby helps to deconstruct stereotypes about people with HCV, while simultaneously increasing positive attitudes among audience members with the ultimate aim of improving quality of health care and treatment for people with HCV.

  3. Aeroacoustic Characterization of the NASA Ames Experimental Aero-Physics Branch 32- by 48-Inch Subsonic Wind Tunnel with a 24-Element Phased Microphone Array

    NASA Technical Reports Server (NTRS)

    Costanza, Bryan T.; Horne, William C.; Schery, S. D.; Babb, Alex T.

    2011-01-01

    The Aero-Physics Branch at NASA Ames Research Center utilizes a 32- by 48-inch subsonic wind tunnel for aerodynamics research. The feasibility of acquiring acoustic measurements with a phased microphone array was recently explored. Acoustic characterization of the wind tunnel was carried out with a floor-mounted 24-element array and two ceiling-mounted speakers. The minimum speaker level for accurate level measurement was evaluated for various tunnel speeds up to a Mach number of 0.15 and streamwise speaker locations. A variety of post-processing procedures, including conventional beamforming and deconvolutional processing such as TIDY, were used. The speaker measurements, with and without flow, were used to compare actual versus simulated in-flow speaker calibrations. Data for wind-off speaker sound and wind-on tunnel background noise were found valuable for predicting sound levels for which the speakers were detectable when the wind was on. Speaker sources were detectable 2 - 10 dB below the peak background noise level with conventional data processing. The effectiveness of background noise cross-spectral matrix subtraction was assessed and found to improve the detectability of test sound sources by approximately 10 dB over a wide frequency range.

  4. Vocal Age Disguise: The Role of Fundamental Frequency and Speech Rate and Its Perceived Effects

    PubMed Central

    Skoog Waller, Sara; Eriksson, Mårten

    2016-01-01

    The relationship between vocal characteristics and perceived age is of interest in various contexts, as is the possibility to affect age perception through vocal manipulation. A few examples of such situations are when age is staged by actors, when ear witnesses make age assessments based on vocal cues only or when offenders (e.g., online groomers) disguise their voice to appear younger or older. This paper investigates how speakers spontaneously manipulate two age related vocal characteristics (f0 and speech rate) in attempt to sound younger versus older than their true age, and if the manipulations correspond to actual age related changes in f0 and speech rate (Study 1). Further aims of the paper is to determine how successful vocal age disguise is by asking listeners to estimate the age of generated speech samples (Study 2) and to examine whether or not listeners use f0 and speech rate as cues to perceived age. In Study 1, participants from three age groups (20–25, 40–45, and 60–65 years) agreed to read a short text under three voice conditions. There were 12 speakers in each age group (six women and six men). They used their natural voice in one condition, attempted to sound 20 years younger in another and 20 years older in a third condition. In Study 2, 60 participants (listeners) listened to speech samples from the three voice conditions in Study 1 and estimated the speakers’ age. Each listener was exposed to all three voice conditions. The results from Study 1 indicated that the speakers increased fundamental frequency (f0) and speech rate when attempting to sound younger and decreased f0 and speech rate when attempting to sound older. Study 2 showed that the voice manipulations had an effect in the sought-after direction, although the achieved mean effect was only 3 years, which is far less than the intended effect of 20 years. Moreover, listeners used speech rate, but not f0, as a cue to speaker age. It was concluded that age disguise by voice can be achieved by naïve speakers even though the perceived effect was smaller than intended. PMID:27917144

  5. The influence of language deprivation in early childhood on L2 processing: An ERP comparison of deaf native signers and deaf signers with a delayed language acquisition.

    PubMed

    Skotara, Nils; Salden, Uta; Kügow, Monique; Hänel-Faulhaber, Barbara; Röder, Brigitte

    2012-05-03

    To examine which language function depends on early experience, the present study compared deaf native signers, deaf non-native signers and hearing German native speakers while processing German sentences. The participants watched simple written sentences while event-related potentials (ERPs) were recorded. At the end of each sentence they were asked to judge whether the sentence was correct or not. Two types of violations were introduced in the middle of the sentence: a semantically implausible noun or a violation of subject-verb number agreement. The results showed a similar ERP pattern after semantic violations (an N400 followed by a positivity) in all three groups. After syntactic violations, native German speakers and native signers of German sign language (DGS) with German as second language (L2) showed a left anterior negativity (LAN) followed by a P600, whereas no LAN but a negativity over the right hemisphere instead was found in deaf participants with a delayed onset of first language (L1) acquisition. The P600 of this group had a smaller amplitude and a different scalp distribution as compared to German native speakers. The results of the present study suggest that language deprivation in early childhood alters the cerebral organization of syntactic language processing mechanisms for L2. Semantic language processing instead was unaffected.

  6. The Whorfian time warp: Representing duration through the language hourglass.

    PubMed

    Bylund, Emanuel; Athanasopoulos, Panos

    2017-07-01

    How do humans construct their mental representations of the passage of time? The universalist account claims that abstract concepts like time are universal across humans. In contrast, the linguistic relativity hypothesis holds that speakers of different languages represent duration differently. The precise impact of language on duration representation is, however, unknown. Here, we show that language can have a powerful role in transforming humans' psychophysical experience of time. Contrary to the universalist account, we found language-specific interference in a duration reproduction task, where stimulus duration conflicted with its physical growth. When reproducing duration, Swedish speakers were misled by stimulus length, and Spanish speakers were misled by stimulus size/quantity. These patterns conform to preferred expressions of duration magnitude in these languages (Swedish: long/short time; Spanish: much/small time). Critically, Spanish-Swedish bilinguals performing the task in both languages showed different interference depending on language context. Such shifting behavior within the same individual reveals hitherto undocumented levels of flexibility in time representation. Finally, contrary to the linguistic relativity hypothesis, language interference was confined to difficult discriminations (i.e., when stimuli varied only subtly in duration and growth), and was eliminated when linguistic cues were removed from the task. These results reveal the malleable nature of human time representation as part of a highly adaptive information processing system. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  7. Engaging spaces: Intimate electro-acoustic display in alternative performance venues

    NASA Astrophysics Data System (ADS)

    Bahn, Curtis; Moore, Stephan

    2004-05-01

    In past presentations to the ASA, we have described the design and construction of four generations of unique spherical speakers (multichannel, outward-radiating geodesic speaker arrays) and Sensor-Speaker-Arrays, (SenSAs: combinations of various sensor devices with outward-radiating multichannel speaker arrays). This presentation will detail the ways in which arrays of these speakers have been employed in alternative performance venues-providing presence and intimacy in the performance of electro-acoustic chamber music and sound installation, while engaging natural and unique acoustical qualities of various locations. We will present documentation of the use of multichannel sonic diffusion arrays in small clubs, ``black-box'' theaters, planetariums, and art galleries.

  8. Speaker diarization system on the 2007 NIST rich transcription meeting recognition evaluation

    NASA Astrophysics Data System (ADS)

    Sun, Hanwu; Nwe, Tin Lay; Koh, Eugene Chin Wei; Bin, Ma; Li, Haizhou

    2007-09-01

    This paper presents a speaker diarization system developed at the Institute for Infocomm Research (I2R) for NIST Rich Transcription 2007 (RT-07) evaluation task. We describe in details our primary approaches for the speaker diarization on the Multiple Distant Microphones (MDM) conditions in conference room scenario. Our proposed system consists of six modules: 1). Least-mean squared (NLMS) adaptive filter for the speaker direction estimate via Time Difference of Arrival (TDOA), 2). An initial speaker clustering via two-stage TDOA histogram distribution quantization approach, 3). Multiple microphone speaker data alignment via GCC-PHAT Time Delay Estimate (TDE) among all the distant microphone channel signals, 4). A speaker clustering algorithm based on GMM modeling approach, 5). Non-speech removal via speech/non-speech verification mechanism and, 6). Silence removal via "Double-Layer Windowing"(DLW) method. We achieves error rate of 31.02% on the 2006 Spring (RT-06s) MDM evaluation task and a competitive overall error rate of 15.32% for the NIST Rich Transcription 2007 (RT-07) MDM evaluation task.

  9. Intonation and gender perception: applications for transgender speakers.

    PubMed

    Hancock, Adrienne; Colton, Lindsey; Douglas, Fiacre

    2014-03-01

    Intonation is commonly addressed in voice and communication feminization therapy, yet empirical evidence of gender differences for intonation is scarce and rarely do studies examine how it relates to gender perception of transgender speakers. This study examined intonation of 12 males, 12 females, six female-to-male, and 14 male-to-female transgender speakers describing a Norman Rockwell image. Several intonation measures were compared between biological gender groups, between perceived gender groups, and between male-to-female (MTF) speakers who were perceived as male, female, or ambiguous gender. Speakers with a larger percentage of utterances with upward intonation and a larger utterance semitone range were perceived as female by listeners, despite no significant differences between the actual intonation of the four gender groups. MTF speakers who do not pass as female appear to use less upward and more downward intonations than female and passing MTF speakers. Intonation has potential for use in transgender communication therapy because it can influence perception to some degree. Copyright © 2014 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  10. The association between tobacco, alcohol, and drug use, stress, and depression among uninsured free clinic patients: U.S.-born English speakers, non-U.S.-born English speakers, and Spanish speakers.

    PubMed

    Kamimura, Akiko; Ashby, Jeanie; Tabler, Jennifer; Nourian, Maziar M; Trinh, Ha Ngoc; Chen, Jason; Reel, Justine J

    2017-01-01

    The abuse of substances is a significant public health issue. Perceived stress and depression have been found to be related to the abuse of substances. The purpose of this study is to examine the prevalence of substance use (i.e., alcohol problems, smoking, and drug use) and the association between substance use, perceived stress, and depression among free clinic patients. Patients completed a self-administered survey in 2015 (N = 504). The overall prevalence of substance use among free clinic patients was not high compared to the U.S. general population. U.S.-born English speakers reported a higher prevalence rate of tobacco smoking and drug use than did non-U.S.-born English speakers and Spanish speakers. Alcohol problems and smoking were significantly related to higher levels of perceived stress and depression. Substance use prevention and education should be included in general health education programs. U.S.-born English speakers would need additional attention. Mental health intervention would be essential to prevention and intervention.

  11. Iran: Regional Perspectives and U.S. Policy

    DTIC Science & Technology

    2010-01-13

    unwilling to publicly challenge Iran on the issue because of their economic dependence on or relationships with Iran...of Iran’s neighbors, like Iraq, and poses direct military threats to others, like Israel and Lebanon. It also directly challenges U.S. efforts to...reports that a former speaker of the Iranian parliament and then- aide to Iran’s Supreme Leader had referred to Bahrain as Iran’s 14th province sparked a

  12. Combining Multiple Knowledge Sources for Speech Recognition

    DTIC Science & Technology

    1988-09-15

    Thus, the first is thle to clarify the pronunciationt ( TASSEAJ for the acronym TASA !). best adaptation sentence, the second sentence, whens addled...10 rapid adapltati,,n sen- tenrces, and 15 spell-i,, de phrases. 6101 resource rirailageo lei SPEAKER-DEPENDENT DATABASE sentences were randortily...combining the smoothed phoneme models with the de - system tested on a standard database using two well de . tailed context models. BYBLOS makes maximal use

  13. Storytelling as an age-dependent skill: oral recall of orally presented stories.

    PubMed

    Mergler, N L; Faust, M; Goldstein, M D

    During experiment 1, three taped prose passages read by college student, middle-aged, or old tellers were orally recalled by college students in an incidental memory paradigm. More story units were remembered as the age of the teller increased (r = +.642, p less than .05). Comparison of these results, with prior research using written, as opposed to oral, presentation and recall of these stories, showed no differences in specific story units remembered. Teller age predicted recall on the two "storied" passages. These passages elicited more favorable comments from listeners when read by older tellers. The third, descriptive passage was less favorably regarded by listeners hearing older tellers. During experiment 2, taped storied passages read by middle-aged tellers were falsely attributed to young, middle-aged, or old persons before the college students listened. Incidental recall did not show an age of teller effect in this case, but the listener's evaluation of the speaker exhibited age-dependent stereotypes. It was concluded that 1) physical qualities of older voices lead to more effective oral transmission; 2) that one expects to receive certain types of oral information from older persons; and 3) that a mismatch between physical vocal quality and age attribution effects evaluation of the speaker, not recall of the information.

  14. A Study on Metadiscoursive Interaction in the MA Theses of the Native Speakers of English and the Turkish Speakers of English

    ERIC Educational Resources Information Center

    Köroglu, Zehra; Tüm, Gülden

    2017-01-01

    This study has been conducted to evaluate the TM usage in the MA theses written by the native speakers (NSs) of English and the Turkish speakers (TSs) of English. The purpose is to compare the TM usage in the introduction, results and discussion, and conclusion sections by both groups' randomly selected MA theses in the field of ELT between the…

  15. Improving the Effectiveness of Speaker Verification Domain Adaptation With Inadequate In-Domain Data

    DTIC Science & Technology

    2017-08-20

    Improving the Effectiveness of Speaker Verification Domain Adaptation With Inadequate In-Domain Data Bengt J. Borgström1, Elliot Singer1, Douglas...ll.mit.edu.edu, dar@ll.mit.edu, es@ll.mit.edu, omid.sadjadi@nist.gov Abstract This paper addresses speaker verification domain adaptation with...contain speakers with low channel diversity. Existing domain adaptation methods are reviewed, and their shortcomings are discussed. We derive an

  16. Mortality inequality in two native population groups.

    PubMed

    Saarela, Jan; Finnäs, Fjalar

    2005-11-01

    A sample of people aged 40-67 years, taken from a longitudinal register compiled by Statistics Finland, is used to analyse mortality differences between Swedish speakers and Finnish speakers in Finland. Finnish speakers are known to have higher death rates than Swedish speakers. The purpose is to explore whether labour-market experience and partnership status, treated as proxies for measures of variation in health-related characteristics, are related to the mortality differential. Persons who are single, disability pensioners, and those having experienced unemployment are found to have substantially higher death rates than those with a partner and employed persons. Swedish speakers have a more favourable distribution on both variables, which thus notably helps to reduce the Finnish-Swedish mortality gradient. A conclusion from this study is that future analyses on the topic should focus on mechanisms that bring a greater proportion of Finnish speakers into the groups with poor health or supposed unhealthy behaviour.

  17. How Psychological Stress Affects Emotional Prosody.

    PubMed

    Paulmann, Silke; Furnes, Desire; Bøkenes, Anne Ming; Cozzolino, Philip J

    2016-01-01

    We explored how experimentally induced psychological stress affects the production and recognition of vocal emotions. In Study 1a, we demonstrate that sentences spoken by stressed speakers are judged by naïve listeners as sounding more stressed than sentences uttered by non-stressed speakers. In Study 1b, negative emotions produced by stressed speakers are generally less well recognized than the same emotions produced by non-stressed speakers. Multiple mediation analyses suggest this poorer recognition of negative stimuli was due to a mismatch between the variation of volume voiced by speakers and the range of volume expected by listeners. Together, this suggests that the stress level of the speaker affects judgments made by the receiver. In Study 2, we demonstrate that participants who were induced with a feeling of stress before carrying out an emotional prosody recognition task performed worse than non-stressed participants. Overall, findings suggest detrimental effects of induced stress on interpersonal sensitivity.

  18. In the eye of the beholder: eye contact increases resistance to persuasion.

    PubMed

    Chen, Frances S; Minson, Julia A; Schöne, Maren; Heinrichs, Markus

    2013-11-01

    Popular belief holds that eye contact increases the success of persuasive communication, and prior research suggests that speakers who direct their gaze more toward their listeners are perceived as more persuasive. In contrast, we demonstrate that more eye contact between the listener and speaker during persuasive communication predicts less attitude change in the direction advocated. In Study 1, participants freely watched videos of speakers expressing various views on controversial sociopolitical issues. Greater direct gaze at the speaker's eyes was associated with less attitude change in the direction advocated by the speaker. In Study 2, we instructed participants to look at either the eyes or the mouths of speakers presenting arguments counter to participants' own attitudes. Intentionally maintaining direct eye contact led to less persuasion than did gazing at the mouth. These findings suggest that efforts at increasing eye contact may be counterproductive across a variety of persuasion contexts.

  19. How Psychological Stress Affects Emotional Prosody

    PubMed Central

    Paulmann, Silke; Furnes, Desire; Bøkenes, Anne Ming; Cozzolino, Philip J.

    2016-01-01

    We explored how experimentally induced psychological stress affects the production and recognition of vocal emotions. In Study 1a, we demonstrate that sentences spoken by stressed speakers are judged by naïve listeners as sounding more stressed than sentences uttered by non-stressed speakers. In Study 1b, negative emotions produced by stressed speakers are generally less well recognized than the same emotions produced by non-stressed speakers. Multiple mediation analyses suggest this poorer recognition of negative stimuli was due to a mismatch between the variation of volume voiced by speakers and the range of volume expected by listeners. Together, this suggests that the stress level of the speaker affects judgments made by the receiver. In Study 2, we demonstrate that participants who were induced with a feeling of stress before carrying out an emotional prosody recognition task performed worse than non-stressed participants. Overall, findings suggest detrimental effects of induced stress on interpersonal sensitivity. PMID:27802287

  20. Don't Underestimate the Benefits of Being Misunderstood.

    PubMed

    Gibson, Edward; Tan, Caitlin; Futrell, Richard; Mahowald, Kyle; Konieczny, Lars; Hemforth, Barbara; Fedorenko, Evelina

    2017-06-01

    Being a nonnative speaker of a language poses challenges. Individuals often feel embarrassed by the errors they make when talking in their second language. However, here we report an advantage of being a nonnative speaker: Native speakers give foreign-accented speakers the benefit of the doubt when interpreting their utterances; as a result, apparently implausible utterances are more likely to be interpreted in a plausible way when delivered in a foreign than in a native accent. Across three replicated experiments, we demonstrated that native English speakers are more likely to interpret implausible utterances, such as "the mother gave the candle the daughter," as similar plausible utterances ("the mother gave the candle to the daughter") when the speaker has a foreign accent. This result follows from the general model of language interpretation in a noisy channel, under the hypothesis that listeners assume a higher error rate in foreign-accented than in nonaccented speech.

  1. Rhythmic patterning in Malaysian and Singapore English.

    PubMed

    Tan, Rachel Siew Kuang; Low, Ee-Ling

    2014-06-01

    Previous work on the rhythm of Malaysian English has been based on impressionistic observations. This paper utilizes acoustic analysis to measure the rhythmic patterns of Malaysian English. Recordings of the read speech and spontaneous speech of 10 Malaysian English speakers were analyzed and compared with recordings of an equivalent sample of Singaporean English speakers. Analysis was done using two rhythmic indexes, the PVI and VarcoV. It was found that although the rhythm of read speech of the Singaporean speakers was syllable-based as described by previous studies, the rhythm of the Malaysian speakers was even more syllable-based. Analysis of the syllables in specific utterances showed that Malaysian speakers did not reduce vowels as much as Singaporean speakers in cases of syllables in utterances. Results of the spontaneous speech confirmed the findings for the read speech; that is, the same rhythmic patterning was found which normally triggers vowel reductions.

  2. Speakers of different languages process the visual world differently.

    PubMed

    Chabal, Sarah; Marian, Viorica

    2015-06-01

    Language and vision are highly interactive. Here we show that people activate language when they perceive the visual world, and that this language information impacts how speakers of different languages focus their attention. For example, when searching for an item (e.g., clock) in the same visual display, English and Spanish speakers look at different objects. Whereas English speakers searching for the clock also look at a cloud, Spanish speakers searching for the clock also look at a gift, because the Spanish names for gift (regalo) and clock (reloj) overlap phonologically. These different looking patterns emerge despite an absence of direct language input, showing that linguistic information is automatically activated by visual scene processing. We conclude that the varying linguistic information available to speakers of different languages affects visual perception, leading to differences in how the visual world is processed. (c) 2015 APA, all rights reserved).

  3. The Basis of the Syllable Hierarchy: Articulatory Pressures or Universal Phonological Constraints?

    PubMed

    Zhao, Xu; Berent, Iris

    2018-02-01

    Across languages, certain syllable types are systematically preferred to others (e.g., [Formula: see text] lbif, where [Formula: see text] indicates a preference). Previous research has shown that these preferences are active in the brains of individual speakers, they are evident even when none of these syllable types exists in participants' language, and even when the stimuli are presented in print. These results suggest that the syllable hierarchy cannot be reduced to either lexical or auditory/phonetic pressures. Here, we examine whether the syllable hierarchy is due to articulatory pressures. According to the motor embodiment view, the perception of a linguistic stimulus requires simulating its production; dispreferred syllables (e.g., lbif) are universally disliked because their production is harder to simulate. To address this possibility, we assessed syllable preferences while articulation was mechanically suppressed. Our four experiments each found significant effects of suppression. Remarkably, people remained sensitive to the syllable hierarchy regardless of suppression. Specifically, results with auditory materials (Experiments 1-2) showed strong effects of syllable structure irrespective of suppression. Moreover, syllable structure uniquely accounted for listeners' behavior even when controlling for several phonetic characteristics of our auditory materials. Results with printed stimuli (Experiments 3-4) were more complex, as participants in these experiments relied on both phonological and graphemic information. Nonetheless, readers were sensitive to most of the syllable hierarchy (e.g., [Formula: see text]), and these preferences emerged when articulation was suppressed, and even when the statistical properties of our materials were controlled via a regression analysis. Together, these findings indicate that speakers possess broad grammatical preferences that are irreducible to either sensory or motor factors.

  4. Processing ser and estar to locate objects and events: An ERP study with L2 speakers of Spanish.

    PubMed

    Dussias, Paola E; Contemori, Carla; Román, Patricia

    2014-01-01

    In Spanish locative constructions, a different form of the copula is selected in relation to the semantic properties of the grammatical subject: sentences that locate objects require estar while those that locate events require ser (both translated in English as 'to be'). In an ERP study, we examined whether second language (L2) speakers of Spanish are sensitive to the selectional restrictions that the different types of subjects impose on the choice of the two copulas. Twenty-four native speakers of Spanish and two groups of L2 Spanish speakers (24 beginners and 18 advanced speakers) were recruited to investigate the processing of 'object/event + estar/ser ' permutations. Participants provided grammaticality judgments on correct (object + estar ; event + ser ) and incorrect (object + ser ; event + estar ) sentences while their brain activity was recorded. In line with previous studies (Leone-Fernández, Molinaro, Carreiras, & Barber, 2012; Sera, Gathje, & Pintado, 1999), the results of the grammaticality judgment for the native speakers showed that participants correctly accepted object + estar and event + ser constructions. In addition, while 'object + ser ' constructions were considered grossly ungrammatical, 'event + estar ' combinations were perceived as unacceptable to a lesser degree. For these same participants, ERP recording time-locked to the onset of the critical word ' en ' showed a larger P600 for the ser predicates when the subject was an object than when it was an event (*La silla es en la cocina vs. La fiesta es en la cocina). This P600 effect is consistent with syntactic repair of the defining predicate when it does not fit with the adequate semantic properties of the subject. For estar predicates (La silla está en la cocina vs. *La fiesta está en la cocina), the findings showed a central-frontal negativity between 500-700 ms. Grammaticality judgment data for the L2 speakers of Spanish showed that beginners were significantly less accurate than native speakers in all conditions, while the advanced speakers only differed from the natives in the event+ ser and event+ estar conditions. For the ERPs, the beginning learners did not show any effects in the time-windows under analysis. The advanced speakers showed a pattern similar to that of native speakers: (1) a P600 response to 'object + ser ' violation more central and frontally distributed, and (2) a central-frontal negativity between 500-700 ms for 'event + estar ' violation. Findings for the advanced speakers suggest that behavioral methods commonly used to assess grammatical knowledge in the L2 may be underestimating what L2 speakers have actually learned.

  5. Reasoning and Comprehension Processes of Linguistic Minority Persons Learning from Text

    DTIC Science & Technology

    1989-08-25

    scores for the ESL speakers are typical for this population. Performance on the Test of English as a Foreign Language ( TOEFL ) is the language proficiency...fluctuated around 500 for the past several years. An additional 7 ESL students reported scores on the Test of English as a Foreign Language ( TOEFL ) and 2...students reported both SAT and TOEFL scores. The mean TOEFL was 564.7, with scores ranging from 510 to 630. 0 The mean TOEFL score is representative of

  6. Knowledge of Connectors as Cohesion in Text: A Comparative Study of Native English and ESL (English as a Second Language) Speakers

    DTIC Science & Technology

    1989-08-18

    for public release; distribution 2b. DECLASSIFICATION / DOWNGRADING SCHEDULE ufnl i mi ted. 4. PERFORMING ORGANIZATION REPORT NUMBER(S) S. MONITORING... ORGANIZATION REPORT NUMBER(S) 6a. NAME OF PERFORMING ORGANIZATION 6b OFFICE SYMBOL 7a. NAME OF MONITORING ORGANIZATION University of California (if...OF FUNDING/SPONSORING 8b OFFICE SYMBOL 9 PROCUREMENT INSTRUMENT IDENTIFICATION NUMBER ORGANIZATION (If applicable) N00014-85-K0562 8c. ADDRESS (City

  7. Reasoning about knowledge: Children's evaluations of generality and verifiability.

    PubMed

    Koenig, Melissa A; Cole, Caitlin A; Meyer, Meredith; Ridge, Katherine E; Kushnir, Tamar; Gelman, Susan A

    2015-12-01

    In a series of experiments, we examined 3- to 8-year-old children's (N=223) and adults' (N=32) use of two properties of testimony to estimate a speaker's knowledge: generality and verifiability. Participants were presented with a "Generic speaker" who made a series of 4 general claims about "pangolins" (a novel animal kind), and a "Specific speaker" who made a series of 4 specific claims about "this pangolin" as an individual. To investigate the role of verifiability, we systematically varied whether the claim referred to a perceptually-obvious feature visible in a picture (e.g., "has a pointy nose") or a non-evident feature that was not visible (e.g., "sleeps in a hollow tree"). Three main findings emerged: (1) young children showed a pronounced reliance on verifiability that decreased with age. Three-year-old children were especially prone to credit knowledge to speakers who made verifiable claims, whereas 7- to 8-year-olds and adults credited knowledge to generic speakers regardless of whether the claims were verifiable; (2) children's attributions of knowledge to generic speakers was not detectable until age 5, and only when those claims were also verifiable; (3) children often generalized speakers' knowledge outside of the pangolin domain, indicating a belief that a person's knowledge about pangolins likely extends to new facts. Findings indicate that young children may be inclined to doubt speakers who make claims they cannot verify themselves, as well as a developmentally increasing appreciation for speakers who make general claims. Copyright © 2015 Elsevier Inc. All rights reserved.

  8. Why We Serve - U.S. Department of Defense Official Website

    Science.gov Websites

    described by a soldier, sailor, airman or Marine who lives it. Story HOW TO HOST A SPEAKER Organizations other organizations. Speakers Photos MEET THE SPEAKERS January 2008 Army Major Lisa L. Carter Navy

  9. Formant transitions in the fluent speech of Farsi-speaking people who stutter.

    PubMed

    Dehqan, Ali; Yadegari, Fariba; Blomgren, Michael; Scherer, Ronald C

    2016-06-01

    Second formant (F2) transitions can be used to infer attributes of articulatory transitions. This study compared formant transitions during fluent speech segments of Farsi (Persian) speaking people who stutter and normally fluent Farsi speakers. Ten Iranian males who stutter and 10 normally fluent Iranian males participated. Sixteen different "CVt" tokens were embedded within the phrase "Begu CVt an". Measures included overall F2 transition frequency extents, durations, and derived overall slopes, initial F2 transition slopes at 30ms and 60ms, and speaking rate. (1) Mean overall formant frequency extent was significantly greater in 14 of the 16 CVt tokens for the group of stuttering speakers. (2) Stuttering speakers exhibited significantly longer overall F2 transitions for all 16 tokens compared to the nonstuttering speakers. (3) The overall F2 slopes were similar between the two groups. (4) The stuttering speakers exhibited significantly greater initial F2 transition slopes (positive or negative) for five of the 16 tokens at 30ms and six of the 16 tokens at 60ms. (5) The stuttering group produced a slower syllable rate than the non-stuttering group. During perceptually fluent utterances, the stuttering speakers had greater F2 frequency extents during transitions, took longer to reach vowel steady state, exhibited some evidence of steeper slopes at the beginning of transitions, had overall similar F2 formant slopes, and had slower speaking rates compared to nonstuttering speakers. Findings support the notion of different speech motor timing strategies in stuttering speakers. Findings are likely to be independent of the language spoken. Educational objectives This study compares aspects of F2 formant transitions between 10 stuttering and 10 nonstuttering speakers. Readers will be able to describe: (a) characteristics of formant frequency as a specific acoustic feature used to infer speech movements in stuttering and nonstuttering speakers, (b) two methods of measuring second formant (F2) transitions: the visual criteria method and fixed time criteria method, (c) characteristics of F2 transitions in the fluent speech of stuttering speakers and how those characteristics appear to differ from normally fluent speakers, and (d) possible cross-linguistic effects on acoustic analyses of stuttering. Copyright © 2016 Elsevier Inc. All rights reserved.

  10. Referential first mention in narratives by mildly mentally retarded adults.

    PubMed

    Kernan, K T; Sabsay, S

    1987-01-01

    Referential first mentions in narrative reports of a short film by 40 mildly mentally retarded adults and 20 nonretarded adults were compared. The mentally retarded sample included equal numbers of male and female, and black and white speakers. The mentally retarded speakers made significantly fewer first mentions and significantly more errors in the form of the first mentions than did nonretarded speakers. A pattern of better performance by black males than by other mentally retarded speakers was found. It is suggested that task difficulty and incomplete mastery of the use of definite and indefinite forms for encoding old and new information, rather than some global type of egocentrism, accounted for the poorer performance by mentally retarded speakers.

  11. Entropy Based Classifier Combination for Sentence Segmentation

    DTIC Science & Technology

    2007-01-01

    speaker diarization system to divide the audio data into hypothetical speakers [17...the prosodic feature also includes turn-based features which describe the position of a word in relation to diarization seg- mentation. The speaker ...ro- bust speaker segmentation: the ICSI-SRI fall 2004 diarization system,” in Proc. RT-04F Workshop, 2004. [18] “The rich transcription fall 2003,” http://nist.gov/speech/tests/rt/rt2003/fall/docs/rt03-fall-eval- plan-v9.pdf.

  12. Somatotype and Body Composition of Normal and Dysphonic Adult Speakers.

    PubMed

    Franco, Débora; Fragoso, Isabel; Andrea, Mário; Teles, Júlia; Martins, Fernando

    2017-01-01

    Voice quality provides information about the anatomical characteristics of the speaker. The patterns of somatotype and body composition can provide essential knowledge to characterize the individuality of voice quality. The aim of this study was to verify if there were significant differences in somatotype and body composition between normal and dysphonic speakers. Cross-sectional study. Anthropometric measurements were taken of a sample of 72 adult participants (40 normal speakers and 32 dysphonic speakers) according to International Society for the Advancement of Kinanthropometry standards, which allowed the calculation of endomorphism, mesomorphism, ectomorphism components, body density, body mass index, fat mass, percentage fat, and fat-free mass. Perception and acoustic evaluations as well as nasoendoscopy were used to assign speakers into normal or dysphonic groups. There were no significant differences between normal and dysphonic speakers in the mean somatotype attitudinal distance and somatotype dispersion distance (in spite of marginally significant differences [P < 0.10] in somatotype attitudinal distance and somatotype dispersion distance between groups) and in the mean vector of the somatotype components. Furthermore, no significant differences were found between groups concerning the mean of percentage fat, fat mass, fat-free mass, body density, and body mass index after controlling by sex. The findings suggested no significant differences in the somatotype and body composition variables, between normal and dysphonic speakers. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  13. Strength of German accent under altered auditory feedback

    PubMed Central

    HOWELL, PETER; DWORZYNSKI, KATHARINA

    2007-01-01

    Borden’s (1979, 1980) hypothesis that speakers with vulnerable speech systems rely more heavily on feedback monitoring than do speakers with less vulnerable systems was investigated. The second language (L2) of a speaker is vulnerable, in comparison with the native language, so alteration to feedback should have a detrimental effect on it, according to this hypothesis. Here, we specifically examined whether altered auditory feedback has an effect on accent strength when speakers speak L2. There were three stages in the experiment. First, 6 German speakers who were fluent in English (their L2) were recorded under six conditions—normal listening, amplified voice level, voice shifted in frequency, delayed auditory feedback, and slowed and accelerated speech rate conditions. Second, judges were trained to rate accent strength. Training was assessed by whether it was successful in separating German speakers speaking English from native English speakers, also speaking English. In the final stage, the judges ranked recordings of each speaker from the first stage as to increasing strength of German accent. The results show that accents were more pronounced under frequency-shifted and delayed auditory feedback conditions than under normal or amplified feedback conditions. Control tests were done to ensure that listeners were judging accent, rather than fluency changes caused by altered auditory feedback. The findings are discussed in terms of Borden’s hypothesis and other accounts about why altered auditory feedback disrupts speech control. PMID:11414137

  14. Political skill: explaining the effects of nonnative accent on managerial hiring and entrepreneurial investment decisions.

    PubMed

    Huang, Laura; Frideger, Marcia; Pearce, Jone L

    2013-11-01

    We propose and test a new theory explaining glass-ceiling bias against nonnative speakers as driven by perceptions that nonnative speakers have weak political skill. Although nonnative accent is a complex signal, its effects on assessments of the speakers' political skill are something that speakers can actively mitigate; this makes it an important bias to understand. In Study 1, White and Asian nonnative speakers using the same scripted responses as native speakers were found to be significantly less likely to be recommended for a middle-management position, and this bias was fully mediated by assessments of their political skill. The alternative explanations of race, communication skill, and collaborative skill were nonsignificant. In Study 2, entrepreneurial start-up pitches from national high-technology, new-venture funding competitions were shown to experienced executive MBA students. Nonnative speakers were found to have a significantly lower likelihood of receiving new-venture funding, and this was fully mediated by the coders' assessments of their political skill. The entrepreneurs' race, communication skill, and collaborative skill had no effect. We discuss the value of empirically testing various posited reasons for glass-ceiling biases, how the importance and ambiguity of political skill for executive success serve as an ostensibly meritocratic cover for nonnative speaker bias, and other theoretical and practical implications of this work. (c) 2013 APA, all rights reserved.

  15. Aphasia in Persian: Implications for cognitive models of lexical processing.

    PubMed

    Bakhtiar, Mehdi; Jafary, Reyhane; Weekes, Brendan S

    2017-09-01

    Current models of oral reading assume that different routes (sublexical, lexical, and semantic) mediate oral reading performance and reliance on different routes during oral reading depends on the characteristics of print to sound mappings. Studies of single cases of acquired dyslexia in aphasia have contributed to the development of such models by revealing patterns of double dissociation in object naming and oral reading skill that follow brain damage in Indo-European and Sino-Tibetan languages. Print to sound mapping in Persian varies in transparency because orthography to phonology translation depends uniquely on the presence or absence of vowel letters in print. Here a hypothesis is tested that oral reading in Persian requires a semantic reading pathway that is independent of a direct non-semantic reading pathway, by investigating whether Persian speakers with aphasia show selective impairments to object naming and reading aloud. A sample of 21 Persian speakers with aphasia ranging in age from 18 to 77 (mean = 53, SD = 16.9) was asked to name a same set of 200 objects and to read aloud the printed names of these objects in different sessions. As an additional measure of sublexical reading, patients were asked to read aloud 30 non-word stimuli. Results showed that oral reading is significantly more preserved than object naming in Persian speakers with aphasia. However, more preserved object naming than oral reading was also observed in some cases. There was a moderate positive correlation between picture naming and oral reading success (p < .05). Mixed-effects logistic regression revealed that word frequency, age of acquisition and imageability predict success across both tasks and there is an interaction between these variables and orthographic transparency in oral reading. Furthermore, opaque words were read less accurately than transparent words. The results reveal different patterns of acquired dyslexia in some cases that closely resemble phonological, deep, and surface dyslexia in other scripts - reported here in Persian for the first time. © 2016 The British Psychological Society.

  16. The role of temporal speech cues in facilitating the fluency of adults who stutter.

    PubMed

    Park, Jin; Logan, Kenneth J

    2015-12-01

    Adults who stutter speak more fluently during choral speech contexts than they do during solo speech contexts. The underlying mechanisms for this effect remain unclear, however. In this study, we examined the extent to which the choral speech effect depended on presentation of intact temporal speech cues. We also examined whether speakers who stutter followed choral signals more closely than typical speakers did. 8 adults who stuttered and 8 adults who did not stutter read 60 sentences aloud during a solo speaking condition and three choral speaking conditions (240 total sentences), two of which featured either temporally altered or indeterminate word duration patterns. Effects of these manipulations on speech fluency, rate, and temporal entrainment with the choral speech signal were assessed. Adults who stutter spoke more fluently in all choral speaking conditions than they did when speaking solo. They also spoke slower and exhibited closer temporal entrainment with the choral signal during the mid- to late-stages of sentence production than the adults who did not stutter. Both groups entrained more closely with unaltered choral signals than they did with altered choral signals. Findings suggest that adults who stutter make greater use of speech-related information in choral signals when talking than adults with typical fluency do. The presence of fluency facilitation during temporally altered choral speech and conversation babble, however, suggests that temporal/gestural cueing alone cannot account for fluency facilitation in speakers who stutter. Other potential fluency enhancing mechanisms are discussed. The reader will be able to (a) summarize competing views on stuttering as a speech timing disorder, (b) describe the extent to which adults who stutter depend on an accurate rendering of temporal information in order to benefit from choral speech, and (c) discuss possible explanations for fluency facilitation in the presence of inaccurate or indeterminate temporal cues. Copyright © 2015 Elsevier Inc. All rights reserved.

  17. Auditory Perceptual Abilities Are Associated with Specific Auditory Experience

    PubMed Central

    Zaltz, Yael; Globerson, Eitan; Amir, Noam

    2017-01-01

    The extent to which auditory experience can shape general auditory perceptual abilities is still under constant debate. Some studies show that specific auditory expertise may have a general effect on auditory perceptual abilities, while others show a more limited influence, exhibited only in a relatively narrow range associated with the area of expertise. The current study addresses this issue by examining experience-dependent enhancement in perceptual abilities in the auditory domain. Three experiments were performed. In the first experiment, 12 pop and rock musicians and 15 non-musicians were tested in frequency discrimination (DLF), intensity discrimination, spectrum discrimination (DLS), and time discrimination (DLT). Results showed significant superiority of the musician group only for the DLF and DLT tasks, illuminating enhanced perceptual skills in the key features of pop music, in which miniscule changes in amplitude and spectrum are not critical to performance. The next two experiments attempted to differentiate between generalization and specificity in the influence of auditory experience, by comparing subgroups of specialists. First, seven guitar players and eight percussionists were tested in the DLF and DLT tasks that were found superior for musicians. Results showed superior abilities on the DLF task for guitar players, though no difference between the groups in DLT, demonstrating some dependency of auditory learning on the specific area of expertise. Subsequently, a third experiment was conducted, testing a possible influence of vowel density in native language on auditory perceptual abilities. Ten native speakers of German (a language characterized by a dense vowel system of 14 vowels), and 10 native speakers of Hebrew (characterized by a sparse vowel system of five vowels), were tested in a formant discrimination task. This is the linguistic equivalent of a DLS task. Results showed that German speakers had superior formant discrimination, demonstrating highly specific effects for auditory linguistic experience as well. Overall, results suggest that auditory superiority is associated with the specific auditory exposure. PMID:29238318

  18. On how the brain decodes vocal cues about speaker confidence.

    PubMed

    Jiang, Xiaoming; Pell, Marc D

    2015-05-01

    In speech communication, listeners must accurately decode vocal cues that refer to the speaker's mental state, such as their confidence or 'feeling of knowing'. However, the time course and neural mechanisms associated with online inferences about speaker confidence are unclear. Here, we used event-related potentials (ERPs) to examine the temporal neural dynamics underlying a listener's ability to infer speaker confidence from vocal cues during speech processing. We recorded listeners' real-time brain responses while they evaluated statements wherein the speaker's tone of voice conveyed one of three levels of confidence (confident, close-to-confident, unconfident) or were spoken in a neutral manner. Neural responses time-locked to event onset show that the perceived level of speaker confidence could be differentiated at distinct time points during speech processing: unconfident expressions elicited a weaker P2 than all other expressions of confidence (or neutral-intending utterances), whereas close-to-confident expressions elicited a reduced negative response in the 330-500 msec and 550-740 msec time window. Neutral-intending expressions, which were also perceived as relatively confident, elicited a more delayed, larger sustained positivity than all other expressions in the 980-1270 msec window for this task. These findings provide the first piece of evidence of how quickly the brain responds to vocal cues signifying the extent of a speaker's confidence during online speech comprehension; first, a rough dissociation between unconfident and confident voices occurs as early as 200 msec after speech onset. At a later stage, further differentiation of the exact level of speaker confidence (i.e., close-to-confident, very confident) is evaluated via an inferential system to determine the speaker's meaning under current task settings. These findings extend three-stage models of how vocal emotion cues are processed in speech comprehension (e.g., Schirmer & Kotz, 2006) by revealing how a speaker's mental state (i.e., feeling of knowing) is simultaneously inferred from vocal expressions. Copyright © 2015 Elsevier Ltd. All rights reserved.

  19. Designing, Modeling, Constructing, and Testing a Flat Panel Speaker and Sound Diffuser for a Simulator

    NASA Technical Reports Server (NTRS)

    Dillon, Christina

    2013-01-01

    The goal of this project was to design, model, build, and test a flat panel speaker and frame for a spherical dome structure being made into a simulator. The simulator will be a test bed for evaluating an immersive environment for human interfaces. This project focused on the loud speakers and a sound diffuser for the dome. The rest of the team worked on an Ambisonics 3D sound system, video projection system, and multi-direction treadmill to create the most realistic scene possible. The main programs utilized in this project, were Pro-E and COMSOL. Pro-E was used for creating detailed figures for the fabrication of a frame that held a flat panel loud speaker. The loud speaker was made from a thin sheet of Plexiglas and 4 acoustic exciters. COMSOL, a multiphysics finite analysis simulator, was used to model and evaluate all stages of the loud speaker, frame, and sound diffuser. Acoustical testing measurements were utilized to create polar plots from the working prototype which were then compared to the COMSOL simulations to select the optimal design for the dome. The final goal of the project was to install the flat panel loud speaker design in addition to a sound diffuser on to the wall of the dome. After running tests in COMSOL on various speaker configurations, including a warped Plexiglas version, the optimal speaker design included a flat piece of Plexiglas with a rounded frame to match the curvature of the dome. Eight of these loud speakers will be mounted into an inch and a half of high performance acoustic insulation, or Thinsulate, that will cover the inside of the dome. The following technical paper discusses these projects and explains the engineering processes used, knowledge gained, and the projected future goals of this project

  20. Perception of speaker size and sex of vowel sounds

    NASA Astrophysics Data System (ADS)

    Smith, David R. R.; Patterson, Roy D.

    2005-04-01

    Glottal-pulse rate (GPR) and vocal-tract length (VTL) are both related to speaker size and sex-however, it is unclear how they interact to determine our perception of speaker size and sex. Experiments were designed to measure the relative contribution of GPR and VTL to judgements of speaker size and sex. Vowels were scaled to represent people with different GPRs and VTLs, including many well beyond the normal population values. In a single interval, two response rating paradigm, listeners judged the size (using a 7-point scale) and sex/age of the speaker (man, woman, boy, or girl) of these scaled vowels. Results from the size-rating experiments show that VTL has a much greater influence upon judgements of speaker size than GPR. Results from the sex-categorization experiments show that judgements of speaker sex are influenced about equally by GPR and VTL for vowels with normal GPR and VTL values. For abnormal combinations of GPR and VTL, where low GPRs are combined with short VTLs, VTL has more influence than GPR in sex judgements. [Work supported by the UK MRC (G9901257) and the German Volkswagen Foundation (VWF 1/79 783).

Top