Sample records for future opening speech

  1. Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces

    PubMed Central

    Bocquelet, Florent; Hueber, Thomas; Girin, Laurent; Savariaux, Christophe; Yvert, Blaise

    2016-01-01

    Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer. PMID:27880768

  2. Emotional recognition from the speech signal for a virtual education agent

    NASA Astrophysics Data System (ADS)

    Tickle, A.; Raghu, S.; Elshaw, M.

    2013-06-01

    This paper explores the extraction of features from the speech wave to perform intelligent emotion recognition. A feature extract tool (openSmile) was used to obtain a baseline set of 998 acoustic features from a set of emotional speech recordings from a microphone. The initial features were reduced to the most important ones so recognition of emotions using a supervised neural network could be performed. Given that the future use of virtual education agents lies with making the agents more interactive, developing agents with the capability to recognise and adapt to the emotional state of humans is an important step.

  3. Securing Our Future. Proceedings of the Conference on Children--Our Future (Kuala Lumpur, Malaysia, November 19-21, 1991).

    ERIC Educational Resources Information Center

    Chiam, Heng Keng, Ed.

    These proceedings report the results of 10 years of ongoing research by the Malaysian Child Development Project to develop and implement measures to study the cognitive, language, and socioemotional development of Malaysian preschool children. Part 1 of the report contains opening speeches delivered by conference organizers and dignitaries. Part 2…

  4. JPRS Report, China

    DTIC Science & Technology

    1993-02-23

    PUBLIC SECURITY New University Military Training Form Introduced [GUANGMING RIBAO 7 Feb] 40 Shijiazhuang Cracks Largest Arms Theft Case ...Taiwan undertook from gradually allowing for freedom of speech in the 1960s, to relaxing press censorship in the 1970s, to opening up the political...Inter- national Futures Company going into operation on 6 February. Its General Manager, Li Qi, said his exchange has joined Malaysia Futures

  5. Fulfilling the Promise of a Sequenced Human Genome – Part II

    ScienceCinema

    Green, Eric

    2018-02-02

    Eric Green, scientific director of the National Human Genome Research Institute (NHGRI), gives the opening keynote speech at the "Sequencing, Finishing, Analysis in the Future" meeting in Santa Fe, NM on May 27, 2009. Part 2 of 2

  6. Fulfilling the Promise of a Sequenced Human Genome – Part I

    ScienceCinema

    Green, Eric

    2018-02-02

    Eric Green, scientific director of the National Human Genome Research Institute (NHGRI), gives the opening keynote speech at the "Sequencing, Finishing, Analysis in the Future" meeting in Santa Fe, NM on May 27, 2009. Part 1 of 2

  7. Nasalance and nasality at experimental velopharyngeal openings in palatal prosthesis: a case study

    PubMed Central

    LIMA-GREGIO, Aveliny Mantovan; MARINO, Viviane Cristina de Castro; PEGORARO-KROOK, Maria Inês; BARBOSA, Plinio Almeida; AFERRI, Homero Carneiro; DUTKA, Jeniffer de Cassia Rillo

    2011-01-01

    The use of prosthetic devices for correction of velopharyngeal insufficiency (VPI) is an alternative treatment for patients with conditions that preclude surgery and for those individuals with a hypofunctional velopharynx (HV) with a poor prognosis for the surgical repair of VPI. Understanding the role and measuring the outcome of prosthetic treatment of velopharyngeal dysfunction requires the use of tools that allow for documenting pre- and post-treatment outcomes. Experimental openings in speech bulbs have been used for simulating VPI in studies documenting changes in aerodynamic, acoustic and kinematics aspects of speech associated with the use of palatal prosthetic devices. The use of nasometry to document changes in speech associated with experimental openings in speech bulbs, however, has not been described in the literature. Objective This single-subject study investigated nasalance and nasality at the presence of experimental openings drilled through the speech bulb of a patient with HV. Material and Methods Nasometric recordings of the word "pato" were obtained under 4 velopharyngeal conditions: no-opening (control condition), no speech bulb, speech bulb with a 20 mm2 opening, and speech bulb with 30 mm2 opening. Five speech-language pathologists performed auditory-perceptual ratings while the subject read an oral passage under all conditions. Results Kruskal-Wallis test showed significant difference among conditions (p=0.0002), with Scheffé post hoc test indicating difference from the no-opening condition. Conclusion The changes in nasalance observed after drilling holes of known sizes in a speech bulb suggest that nasometry reflect changes in transfer of sound energy related to different sizes of velopharyngeal opening. PMID:22230996

  8. Noise and communication: a three-year update.

    PubMed

    Brammer, Anthony J; Laroche, Chantal

    2012-01-01

    Noise is omnipresent and impacts us all in many aspects of daily living. Noise can interfere with communication not only in industrial workplaces, but also in other work settings (e.g. open-plan offices, construction, and mining) and within buildings (e.g. residences, arenas, and schools). The interference of noise with communication can have significant social consequences, especially for persons with hearing loss, and may compromise safety (e.g. failure to perceive auditory warning signals), influence worker productivity and learning in children, affect health (e.g. vocal pathology, noise-induced hearing loss), compromise speech privacy, and impact social participation by the elderly. For workers, attempts have been made to: 1) Better define the auditory performance needed to function effectively and to directly measure these abilities when assessing Auditory Fitness for Duty, 2) design hearing protection devices that can improve speech understanding while offering adequate protection against loud noises, and 3) improve speech privacy in open-plan offices. As the elderly are particularly vulnerable to the effects of noise, an understanding of the interplay between auditory, cognitive, and social factors and its effect on speech communication and social participation is also critical. Classroom acoustics and speech intelligibility in children have also gained renewed interest because of the importance of effective speech comprehension in noise on learning. Finally, substantial work has been made in developing models aimed at better predicting speech intelligibility. Despite progress in various fields, the design of alarm signals continues to lag behind advancements in knowledge. This summary of the last three years' research highlights some of the most recent issues for the workplace, for older adults, and for children, as well as the effectiveness of warning sounds and models for predicting speech intelligibility. Suggestions for future work are also discussed.

  9. Remarks by the President at Princeton University Commencement Address.

    ERIC Educational Resources Information Center

    Clinton, William

    This speech by President William Clinton addresses the 1996 graduating class of Princeton University on the occasion of Princeton's 250th anniversary. After his opening remarks, the President touches on the importance of the graduating seniors' education as well as the importance of education in general to the nation's future. He compares the…

  10. Improved Speech Coding Based on Open-Loop Parameter Estimation

    NASA Technical Reports Server (NTRS)

    Juang, Jer-Nan; Chen, Ya-Chin; Longman, Richard W.

    2000-01-01

    A nonlinear optimization algorithm for linear predictive speech coding was developed early that not only optimizes the linear model coefficients for the open loop predictor, but does the optimization including the effects of quantization of the transmitted residual. It also simultaneously optimizes the quantization levels used for each speech segment. In this paper, we present an improved method for initialization of this nonlinear algorithm, and demonstrate substantial improvements in performance. In addition, the new procedure produces monotonically improving speech quality with increasing numbers of bits used in the transmitted error residual. Examples of speech encoding and decoding are given for 8 speech segments and signal to noise levels as high as 47 dB are produced. As in typical linear predictive coding, the optimization is done on the open loop speech analysis model. Here we demonstrate that minimizing the error of the closed loop speech reconstruction, instead of the simpler open loop optimization, is likely to produce negligible improvement in speech quality. The examples suggest that the algorithm here is close to giving the best performance obtainable from a linear model, for the chosen order with the chosen number of bits for the codebook.

  11. A Multimodal Dialog System for Language Assessment: Current State and Future Directions. Research Report. ETS RR-17-21

    ERIC Educational Resources Information Center

    Suendermann-Oeft, David; Ramanarayanan, Vikram; Yu, Zhou; Qian, Yao; Evanini, Keelan; Lange, Patrick; Wang, Xinhao; Zechner, Klaus

    2017-01-01

    We present work in progress on a multimodal dialog system for English language assessment using a modular cloud-based architecture adhering to open industry standards. Among the modules being developed for the system, multiple modules heavily exploit machine learning techniques, including speech recognition, spoken language proficiency rating,…

  12. Temporal Sensitivity Measured Shortly After Cochlear Implantation Predicts 6-Month Speech Recognition Outcome.

    PubMed

    Erb, Julia; Ludwig, Alexandra Annemarie; Kunke, Dunja; Fuchs, Michael; Obleser, Jonas

    2018-04-24

    Psychoacoustic tests assessed shortly after cochlear implantation are useful predictors of the rehabilitative speech outcome. While largely independent, both spectral and temporal resolution tests are important to provide an accurate prediction of speech recognition. However, rapid tests of temporal sensitivity are currently lacking. Here, we propose a simple amplitude modulation rate discrimination (AMRD) paradigm that is validated by predicting future speech recognition in adult cochlear implant (CI) patients. In 34 newly implanted patients, we used an adaptive AMRD paradigm, where broadband noise was modulated at the speech-relevant rate of ~4 Hz. In a longitudinal study, speech recognition in quiet was assessed using the closed-set Freiburger number test shortly after cochlear implantation (t0) as well as the open-set Freiburger monosyllabic word test 6 months later (t6). Both AMRD thresholds at t0 (r = -0.51) and speech recognition scores at t0 (r = 0.56) predicted speech recognition scores at t6. However, AMRD and speech recognition at t0 were uncorrelated, suggesting that those measures capture partially distinct perceptual abilities. A multiple regression model predicting 6-month speech recognition outcome with deafness duration and speech recognition at t0 improved from adjusted R = 0.30 to adjusted R = 0.44 when AMRD threshold was added as a predictor. These findings identify AMRD thresholds as a reliable, nonredundant predictor above and beyond established speech tests for CI outcome. This AMRD test could potentially be developed into a rapid clinical temporal-resolution test to be integrated into the postoperative test battery to improve the reliability of speech outcome prognosis.

  13. A laboratory study for assessing speech privacy in a simulated open-plan office.

    PubMed

    Lee, P J; Jeon, J Y

    2014-06-01

    The aim of this study is to assess speech privacy in open-plan office using two recently introduced single-number quantities: the spatial decay rate of speech, DL(2,S) [dB], and the A-weighted sound pressure level of speech at a distance of 4 m, L(p,A,S,4) m [dB]. Open-plan offices were modeled using a DL(2,S) of 4, 8, and 12 dB, and L(p,A,S,4) m was changed in three steps, from 43 to 57 dB.Auditory experiments were conducted at three locations with source–receiver distances of 8, 16, and 24 m, while background noise level was fixed at 30 dBA.A total of 20 subjects were asked to rate the speech intelligibility and listening difficulty of 240 Korean sentences in such surroundings. The speech intelligibility scores were not affected by DL(2,S) or L(p,A,S,4) m at a source–receiver distance of 8 m; however, listening difficulty ratings were significantly changed with increasing DL(2,S) and L(p,A,S,4) m values. At other locations, the influences of DL(2,S) and L(p,A,S,4) m on speech intelligibility and listening difficulty ratings were significant. It was also found that the speech intelligibility scores and listening difficulty ratings were considerably changed with increasing the distraction distance (r(D)). Furthermore, listening difficulty is more sensitive to variations in DL(2,S) and L(p,A,S,4) m than intelligibility scores for sound fields with high speech transmission performances. The recently introduced single-number quantities in the ISO standard, based on the spatial distribution of sound pressure level, were associated with speech privacy in an open-plan office. The results support single-number quantities being suitable to assess speech privacy, mainly at large distances. This new information can be considered when designing open-plan offices and making acoustic guidelines of open-plan offices.

  14. A survey of acoustic conditions in semi-open plan classrooms in the United Kingdom.

    PubMed

    Greenland, Emma E; Shield, Bridget M

    2011-09-01

    This paper reports the results of a large scale, detailed acoustic survey of 42 open plan classrooms of varying design in the UK each of which contained between 2 and 14 teaching areas or classbases. The objective survey procedure, which was designed specifically for use in open plan classrooms, is described. The acoustic measurements relating to speech intelligibility within a classbase, including ambient noise level, intrusive noise level, speech to noise ratio, speech transmission index, and reverberation time, are presented. The effects on speech intelligibility of critical physical design variables, such as the number of classbases within an open plan unit and the selection of acoustic finishes for control of reverberation, are examined. This analysis enables limitations of open plan classrooms to be discussed and acoustic design guidelines to be developed to ensure good listening conditions. The types of teaching activity to provide adequate acoustic conditions, plus the speech intelligibility requirements of younger children, are also discussed. © 2011 Acoustical Society of America

  15. The Development of the Mealings, Demuth, Dillon, and Buchholz Classroom Speech Perception Test

    ERIC Educational Resources Information Center

    Mealings, Kiri T.; Demuth, Katherine; Buchholz, Jörg; Dillon, Harvey

    2015-01-01

    Purpose: Open-plan classroom styles are increasingly being adopted in Australia despite evidence that their high intrusive noise levels adversely affect learning. The aim of this study was to develop a new Australian speech perception task (the Mealings, Demuth, Dillon, and Buchholz Classroom Speech Perception Test) and use it in an open-plan…

  16. Opening Statements and Speeches. Plenary Session. Papers.

    ERIC Educational Resources Information Center

    International Federation of Library Associations, The Hague (Netherlands).

    Official opening statements, organizational reports, and papers on libraries in a technological world, which were presented at the 1983 International Federation of Library Associations (IFLA) conference include: (1) welcoming addresses by Franz Georg Kaltwasser and Mathilde Berghofer-Weichner; (2) opening speeches by Else Granheim (IFLA president)…

  17. The Importance of Quality in the CEU. Proceedings of the Annual Conference of the Council on the Continuing Education Unit (2nd, Memphis, Tennessee, June 12-13, 1980).

    ERIC Educational Resources Information Center

    Medick, Glenn A., Ed.

    These proceedings contain twenty-one presentations made at the Second Annual Conference of the Council on the Continuing Education Unit (CEU). The opening speeches--Introduction to the CEU and Future Directions for a Learning Society--are followed by the panel discussion presentation, The CEU: Can It Withstand Scrutiny?, and two respondent…

  18. Improved Open-Microphone Speech Recognition

    NASA Astrophysics Data System (ADS)

    Abrash, Victor

    2002-12-01

    Many current and future NASA missions make extreme demands on mission personnel both in terms of work load and in performing under difficult environmental conditions. In situations where hands are impeded or needed for other tasks, eyes are busy attending to the environment, or tasks are sufficiently complex that ease of use of the interface becomes critical, spoken natural language dialog systems offer unique input and output modalities that can improve efficiency and safety. They also offer new capabilities that would not otherwise be available. For example, many NASA applications require astronauts to use computers in micro-gravity or while wearing space suits. Under these circumstances, command and control systems that allow users to issue commands or enter data in hands-and eyes-busy situations become critical. Speech recognition technology designed for current commercial applications limits the performance of the open-ended state-of-the-art dialog systems being developed at NASA. For example, today's recognition systems typically listen to user input only during short segments of the dialog, and user input outside of these short time windows is lost. Mistakes detecting the start and end times of user utterances can lead to mistakes in the recognition output, and the dialog system as a whole has no way to recover from this, or any other, recognition error. Systems also often require the user to signal when that user is going to speak, which is impractical in a hands-free environment, or only allow a system-initiated dialog requiring the user to speak immediately following a system prompt. In this project, SRI has developed software to enable speech recognition in a hands-free, open-microphone environment, eliminating the need for a push-to-talk button or other signaling mechanism. The software continuously captures a user's speech and makes it available to one or more recognizers. By constantly monitoring and storing the audio stream, it provides the spoken dialog manager extra flexibility to recognize the signal with no audio gaps between recognition requests, as well as to rerecognize portions of the signal, or to rerecognize speech with different grammars, acoustic models, recognizers, start times, and so on. SRI expects that this new open-mic functionality will enable NASA to develop better error-correction mechanisms for spoken dialog systems, and may also enable new interaction strategies.

  19. Improved Open-Microphone Speech Recognition

    NASA Technical Reports Server (NTRS)

    Abrash, Victor

    2002-01-01

    Many current and future NASA missions make extreme demands on mission personnel both in terms of work load and in performing under difficult environmental conditions. In situations where hands are impeded or needed for other tasks, eyes are busy attending to the environment, or tasks are sufficiently complex that ease of use of the interface becomes critical, spoken natural language dialog systems offer unique input and output modalities that can improve efficiency and safety. They also offer new capabilities that would not otherwise be available. For example, many NASA applications require astronauts to use computers in micro-gravity or while wearing space suits. Under these circumstances, command and control systems that allow users to issue commands or enter data in hands-and eyes-busy situations become critical. Speech recognition technology designed for current commercial applications limits the performance of the open-ended state-of-the-art dialog systems being developed at NASA. For example, today's recognition systems typically listen to user input only during short segments of the dialog, and user input outside of these short time windows is lost. Mistakes detecting the start and end times of user utterances can lead to mistakes in the recognition output, and the dialog system as a whole has no way to recover from this, or any other, recognition error. Systems also often require the user to signal when that user is going to speak, which is impractical in a hands-free environment, or only allow a system-initiated dialog requiring the user to speak immediately following a system prompt. In this project, SRI has developed software to enable speech recognition in a hands-free, open-microphone environment, eliminating the need for a push-to-talk button or other signaling mechanism. The software continuously captures a user's speech and makes it available to one or more recognizers. By constantly monitoring and storing the audio stream, it provides the spoken dialog manager extra flexibility to recognize the signal with no audio gaps between recognition requests, as well as to rerecognize portions of the signal, or to rerecognize speech with different grammars, acoustic models, recognizers, start times, and so on. SRI expects that this new open-mic functionality will enable NASA to develop better error-correction mechanisms for spoken dialog systems, and may also enable new interaction strategies.

  20. Women, Clerical Work, and Office Automation: Issues for Research. Report of a Conference Sponsored by the Women's Bureau in Cooperation with the Panel on Technology and Women's Employment National Research Council (Washington, D.C., October 19, 1984).

    ERIC Educational Resources Information Center

    National Academy of Sciences - National Research Council, Washington, DC. Panel on Technology and Women's Employment.

    These proceedings contain presentations (speeches, discussions, papers) from a conference on how office automation is affecting the work lives and employment future of clerical workers. They include a "Welcome to the [National] Academy [of Sciences]" (Roslyn Feldberg), "Opening Remarks" (Lenora Cole Alexander), and "Goals…

  1. International Federation of Library Associations General Conference, Montreal 1982. Official Opening Statements and Speeches. Plenary Session I and II. Papers.

    ERIC Educational Resources Information Center

    International Federation of Library Associations, The Hague (Netherlands).

    Official opening statements and papers on networking and the development of information technology which were presented at the 1982 International Federation of Library Associations (IFLA) conference include: (1) opening speeches by Else Granheim (IFLA president) and Kenneth H. Rogers (UNESCO Representative); (2) "The Importance of Networks…

  2. Spotlight on Speech Codes 2012: The State of Free Speech on Our Nation's Campuses

    ERIC Educational Resources Information Center

    Foundation for Individual Rights in Education (NJ1), 2012

    2012-01-01

    The U.S. Supreme Court has called America's colleges and universities "vital centers for the Nation's intellectual life," but the reality today is that many of these institutions severely restrict free speech and open debate. Speech codes--policies prohibiting student and faculty speech that would, outside the bounds of campus, be…

  3. Collaborative Signaling of Informational Structures by Dynamic Speech Rate.

    ERIC Educational Resources Information Center

    Koiso, Hanae; Shimojima, Atsushi; Katagiri, Yasuhiro

    1998-01-01

    Investigated the functions of dynamic speech rates as contextualization cues in conversational Japanese, examining five spontaneous task-oriented dialogs and analyzing the potential of speech-rate changes in signaling the structure of the information being exchanged. Results found a correlation between speech decelerations and the openings of new…

  4. Visual Speech Primes Open-Set Recognition of Spoken Words

    ERIC Educational Resources Information Center

    Buchwald, Adam B.; Winters, Stephen J.; Pisoni, David B.

    2009-01-01

    Visual speech perception has become a topic of considerable interest to speech researchers. Previous research has demonstrated that perceivers neurally encode and use speech information from the visual modality, and this information has been found to facilitate spoken word recognition in tasks such as lexical decision (Kim, Davis, & Krins,…

  5. Walter Reed Army Medical Center (WRAMC) and Office of Management and Budget (OMB) Circular A-76: Implications for the Future

    DTIC Science & Technology

    2008-08-20

    Opening Statement, The Honorable John O. Marsh, Jr., before the Subcommittee on Military Personnel, House Armed Services Committee, June 26, 2007. 23 Speech...asset-management firm chaired by former Treasury secretary John W. Snow. The company is headed by two former high-ranking executives of KBR, formerly...known as Kellogg Brown & Root. Al Neffgen, IAP’s chief executive, was chief operating officer for a KBR division before joining IAP in 2004. IAP’s

  6. A model of mandibular movements during speech: normative pilot study for the Brazilian Portuguese language.

    PubMed

    Bianchini, Esther M G; de Andrade, Cláudia R F

    2006-07-01

    The precision of speech articulation is related to the possibility and freedom of the mandibular movements, modifying the spaces in order to allow the different articulatory positions of each sound. Electrognathography allows the objective delineation and registration of the mandibular movements, determining the level of opening, translations and velocity of these movements. Its use is a resource that can establish quantitative diagnostic parameters. The aim of this study was to verify the amplitude, velocity and characterization of the mandibular movements during speech using computerized electrognathography. Participants were 40 adults, male and female, with no temporomandibular disorders; with no missing teeth; with no dental occlusion alterations or dentofacial deformities; with no dental prostheses; and with no communication, neurological or cognitive deficits. The mandibular movements were observed during the sequential naming of pictures containing all the phonemes of the Brazilian Portuguese language. The registrations were obtained using electrognathography (BioENG-BioPak system), assessing the spatial position, course and velocity of the mandibular movements. The mean values of velocity were: 88.65 mm/sec during opening and 89.90mm/sec during closing. The mean values of amplitude were: sagittal opening: 12.77 mm, frontal opening: 11.21 mm, protrusion: 1.22 mm; retrusion 5.67 mm; translations to the right: 1.49 mm and to the left: 1.59 mm. The velocity of opening is directly related to that of closing. The amplitude of opening demonstrates a direct correlation with the velocity of opening and closing. All participants presented lateral translations during the course of the jaw. The assessment of speech in normal individuals is characterized by: discreet mandibular movements with an anteroposterior component and lateral translations. This study allowed for the delineation of a profile of the mandibular movements during speech in asymptomatic individuals.

  7. Segregation of Whispered Speech Interleaved with Noise or Speech Maskers

    DTIC Science & Technology

    2011-08-01

    range over which the talker can be heard. Whispered speech is produced by modulating the flow of air through partially open vocal folds. Because the...source of excitation is turbulent air flow , the acoustic characteristics of whispered speech differs from voiced speech [1, 2]. Despite the acoustic...signals provided by cochlear implants. Two studies investigated the segregation of simultaneously presented whispered vowels [7, 8] in a standard

  8. Why do speech and language therapists stay in, leave and (sometimes) return to the National Health Service (NHS)?

    PubMed

    Loan-Clarke, John; Arnold, John; Coombs, Crispin; Bosley, Sara; Martin, Caroline

    2009-01-01

    Research into recruitment, retention and return of speech and language therapists in the National Health Service (NHS) is relatively limited, particularly in respect of understanding the factors that drive employment choice decisions. To identify what factors influence speech and language therapists working in the NHS to stay, and consider leaving, but not do so. To identify what features of the NHS and alternative employers influence speech and language therapists to leave the NHS. To identify why some speech and language therapists return to the NHS after working elsewhere. A total of 516 male and female speech and language therapists, in three distinct groups (NHS stayers, leavers and returners) completed a questionnaire and gave responses to open-ended questions regarding their perceptions of the NHS and other employers. Qualitative data analysis identified reasons why individuals stayed in, left or returned to the NHS employment, and what actions could be taken by management to facilitate retention and return. Stayers value job and pension security; professional development opportunities; the work itself; and professional support. Leavers not involved in childcare left because of workload/pressure/stress; poor pay; and not being able to give good patient care. Returners returned because of flexible hours; work location; professional development; and pension provision. Stayers and returners primarily wish to see more staff in the NHS, whereas leavers would return if there were more flexibility in work arrangements. Returners were particularly hostile towards Agenda for Change. Whilst some preferences appear to require increased resources, others such as reducing bureaucracy and valuing professionals do not. The full impact of Agenda for Change has yet to be established. Predicted excess labour supply of allied health professionals and future structural changes present pressures but also possible opportunities for speech and language therapists.

  9. Long short-term memory for speaker generalization in supervised speech separation

    PubMed Central

    Chen, Jitong; Wang, DeLiang

    2017-01-01

    Speech separation can be formulated as learning to estimate a time-frequency mask from acoustic features extracted from noisy speech. For supervised speech separation, generalization to unseen noises and unseen speakers is a critical issue. Although deep neural networks (DNNs) have been successful in noise-independent speech separation, DNNs are limited in modeling a large number of speakers. To improve speaker generalization, a separation model based on long short-term memory (LSTM) is proposed, which naturally accounts for temporal dynamics of speech. Systematic evaluation shows that the proposed model substantially outperforms a DNN-based model on unseen speakers and unseen noises in terms of objective speech intelligibility. Analyzing LSTM internal representations reveals that LSTM captures long-term speech contexts. It is also found that the LSTM model is more advantageous for low-latency speech separation and it, without future frames, performs better than the DNN model with future frames. The proposed model represents an effective approach for speaker- and noise-independent speech separation. PMID:28679261

  10. Rethinking Adult Education for Development II. Conference Proceedings (Ljubljana, Slovenia, October 6-9, 1993).

    ERIC Educational Resources Information Center

    Svetina, Metka, Ed.

    This book contains papers, reports, and opening and closing speeches presented at a worldwide conference of adult education experts. The document begins with the following: an introduction (Jelenc); a reprinted article, "Time to Collectively Rethink Our Field" (Franklin W. Spikes); and opening speeches by various adult education officials (Jelenc,…

  11. [Acoustic conditions in open plan office - Application of technical measures in a typical room].

    PubMed

    Mikulski, Witold

    2018-03-09

    Noise in open plan offices should not exceed acceptable levels for the hearing protection. Its major negative effects on employees are nuisance and impediment in execution of work. Specific technical solutions should be introduced to provide proper acoustic conditions for work performance. Acoustic evaluation of a typical open plan office was presented in the article published in "Medycyna Pracy" 5/2016. None of the rooms meets all the criteria, therefore, in this article one of the rooms was chosen to apply different technical solutions to check the possibility of reaching proper acoustic conditions. Acoustic effectiveness of those solutions was verified by means of digital simulation. The model was checked by comparing the results of measurements and calculations before using simulation. The analyzis revealed that open plan offices supplemented with signals for masking speech signals can meet all the required criteria. It is relatively easy to reach proper reverberation time (i.e., sound absorption). It is more difficult to reach proper values of evaluation parameters determined from A-weighted sound pressure level (SPLA) of speech. The most difficult is to provide proper values of evaluation parameters determined from speech transmission index (STI). Finally, it is necessary (besides acoustic treatment) to use devices for speech masking. The study proved that it is technically possible to reach proper acoustic condition. Main causes of employees complaints in open plan office are inadequate acoustic work conditions. Therefore, it is necessary to apply specific technical solutions - not only sound absorbing suspended ceiling and high acoustic barriers, but also devices for speech masking. Med Pr 2018;69(2):153-165. This work is available in Open Access model and licensed under a CC BY-NC 3.0 PL license.

  12. Open source OCR framework using mobile devices

    NASA Astrophysics Data System (ADS)

    Zhou, Steven Zhiying; Gilani, Syed Omer; Winkler, Stefan

    2008-02-01

    Mobile phones have evolved from passive one-to-one communication device to powerful handheld computing device. Today most new mobile phones are capable of capturing images, recording video, and browsing internet and do much more. Exciting new social applications are emerging on mobile landscape, like, business card readers, sing detectors and translators. These applications help people quickly gather the information in digital format and interpret them without the need of carrying laptops or tablet PCs. However with all these advancements we find very few open source software available for mobile phones. For instance currently there are many open source OCR engines for desktop platform but, to our knowledge, none are available on mobile platform. Keeping this in perspective we propose a complete text detection and recognition system with speech synthesis ability, using existing desktop technology. In this work we developed a complete OCR framework with subsystems from open source desktop community. This includes a popular open source OCR engine named Tesseract for text detection & recognition and Flite speech synthesis module, for adding text-to-speech ability.

  13. Movement of the velum during speech and singing in classically trained singers.

    PubMed

    Austin, S F

    1997-06-01

    The present study addresses two questions: (a) Is the action and/or posture of the velopharyngeal valve conducive to allow significant resonance during Western tradition classical singing? (b) How do the actions of the velopharyngeal valve observed in this style of singing compare with normal speech? A photodetector system was used to observe the area function of the velopharyngeal port during speech and classical style singing. Identical speech samples were produced by each subject in a normal speaking voice and then in the low, medium, and high singing ranges. Results indicate that in these four singers the velopharyngeal port was closed significantly longer in singing than in speaking samples. The amount of time the velopharyngeal port was opened was greatest in speech and diminished as the singer ascended in pitch. In the high voice condition, little or no opening of the velopharyngeal port was measured.

  14. [Submental island pedicled flap combination with bio-membrane for reconstructing the piercing palate defects].

    PubMed

    Liu, Hanqian; Yu, Huiming; Liu, Jiawu; Fang, Jin; Mao, Chi

    2015-05-01

    To evaluate the clinical outcomes of submental island pedicled flap (SIPF) combination with bio-membrane in reconstructing palate defects after maxillofacial or palatal neoplasm resection. There were 12 patients with squamous cell carcinoma and one patient with adenoid cystic carcinoma. The clinical stages of tumours were II in two patients, III in four patients, IV in six patients (UICC 2002), and one patient with adenoid cystic carcinoma no staged. SIPFs were designed and created, and the tissue sides of the SIPFs were covered with bio-membrane to reconstruct the oral and the nasal sides of the defects respectively. Speech and swallowing functions and opening mouth were evaluated 6 months postoperatively. All flaps survived and no serious complications occurred. Ten patients achieved normal speech, two had intelligible speech, and one was with slurred speech; Nine patients resumed a solid diet, three with a soft diet, and one on a liquid diet. Eight patients recovered normal mouth opening, four emerged minor limitation of mouth opening, and one had serious limitation of mouth opening. SIPF combined with bio-membrane is a safe, simple, and reliable method for reconstruction of piercing palate defect following neoplasm ablation, with satisfactory oral functions.

  15. Speech-like orofacial oscillations in stump-tailed macaque (Macaca arctoides) facial and vocal signals.

    PubMed

    Toyoda, Aru; Maruhashi, Tamaki; Malaivijitnond, Suchinda; Koda, Hiroki

    2017-10-01

    Speech is unique to humans and characterized by facial actions of ∼5 Hz oscillations of lip, mouth or jaw movements. Lip-smacking, a facial display of primates characterized by oscillatory actions involving the vertical opening and closing of the jaw and lips, exhibits stable 5-Hz oscillation patterns, matching that of speech, suggesting that lip-smacking is a precursor of speech. We tested if facial or vocal actions exhibiting the same rate of oscillation are found in wide forms of facial or vocal displays in various social contexts, exhibiting diversity among species. We observed facial and vocal actions of wild stump-tailed macaques (Macaca arctoides), and selected video clips including facial displays (teeth chattering; TC), panting calls, and feeding. Ten open-to-open mouth durations during TC and feeding and five amplitude peak-to-peak durations in panting were analyzed. Facial display (TC) and vocalization (panting) oscillated within 5.74 ± 1.19 and 6.71 ± 2.91 Hz, respectively, similar to the reported lip-smacking of long-tailed macaques and the speech of humans. These results indicated a common mechanism for the central pattern generator underlying orofacial movements, which would evolve to speech. Similar oscillations in panting, which evolved from different muscular control than the orofacial action, suggested the sensory foundations for perceptual saliency particular to 5-Hz rhythms in macaques. This supports the pre-adaptation hypothesis of speech evolution, which states a central pattern generator for 5-Hz facial oscillation and perceptual background tuned to 5-Hz actions existed in common ancestors of macaques and humans, before the emergence of speech. © 2017 Wiley Periodicals, Inc.

  16. Mumbling: Macho or Morphoogy?

    ERIC Educational Resources Information Center

    Weirich, Melanie; Fuchs, Susanne; Simpson, Adrian; Winkler, Ralf; Perrier, Pascal

    2016-01-01

    Purpose: Mumbling as opposed to clear speech is a typical male characteristic in speech and can be the consequence of a small jaw opening. Whereas behavioral reasons have often been offered to explain sex-specific differences with respect to clear speech, the purpose of this study is to investigate a potential anatomical reason for smaller jaw…

  17. Speech Output Technologies in Interventions for Individuals with Autism Spectrum Disorders: A Scoping Review.

    PubMed

    Schlosser, Ralf W; Koul, Rajinder K

    2015-01-01

    The purpose of this scoping review was to (a) map the research evidence on the effectiveness of augmentative and alternative communication (AAC) interventions using speech output technologies (e.g., speech-generating devices, mobile technologies with AAC-specific applications, talking word processors) for individuals with autism spectrum disorders, (b) identify gaps in the existing literature, and (c) posit directions for future research. Outcomes related to speech, language, and communication were considered. A total of 48 studies (47 single case experimental designs and 1 randomized control trial) involving 187 individuals were included. Results were reviewed in terms of three study groupings: (a) studies that evaluated the effectiveness of treatment packages involving speech output, (b) studies comparing one treatment package with speech output to other AAC modalities, and (c) studies comparing the presence with the absence of speech output. The state of the evidence base is discussed and several directions for future research are posited.

  18. Early Postimplant Speech Perception and Language Skills Predict Long-Term Language and Neurocognitive Outcomes Following Pediatric Cochlear Implantation

    PubMed Central

    Kronenberger, William G.; Castellanos, Irina; Pisoni, David B.

    2017-01-01

    Purpose We sought to determine whether speech perception and language skills measured early after cochlear implantation in children who are deaf, and early postimplant growth in speech perception and language skills, predict long-term speech perception, language, and neurocognitive outcomes. Method Thirty-six long-term users of cochlear implants, implanted at an average age of 3.4 years, completed measures of speech perception, language, and executive functioning an average of 14.4 years postimplantation. Speech perception and language skills measured in the 1st and 2nd years postimplantation and open-set word recognition measured in the 3rd and 4th years postimplantation were obtained from a research database in order to assess predictive relations with long-term outcomes. Results Speech perception and language skills at 6 and 18 months postimplantation were correlated with long-term outcomes for language, verbal working memory, and parent-reported executive functioning. Open-set word recognition was correlated with early speech perception and language skills and long-term speech perception and language outcomes. Hierarchical regressions showed that early speech perception and language skills at 6 months postimplantation and growth in these skills from 6 to 18 months both accounted for substantial variance in long-term outcomes for language and verbal working memory that was not explained by conventional demographic and hearing factors. Conclusion Speech perception and language skills measured very early postimplantation, and early postimplant growth in speech perception and language, may be clinically relevant markers of long-term language and neurocognitive outcomes in users of cochlear implants. Supplemental materials https://doi.org/10.23641/asha.5216200 PMID:28724130

  19. Two Methods of Automatic Evaluation of Speech Signal Enhancement Recorded in the Open-Air MRI Environment

    NASA Astrophysics Data System (ADS)

    Přibil, Jiří; Přibilová, Anna; Frollo, Ivan

    2017-12-01

    The paper focuses on two methods of evaluation of successfulness of speech signal enhancement recorded in the open-air magnetic resonance imager during phonation for the 3D human vocal tract modeling. The first approach enables to obtain a comparison based on statistical analysis by ANOVA and hypothesis tests. The second method is based on classification by Gaussian mixture models (GMM). The performed experiments have confirmed that the proposed ANOVA and GMM classifiers for automatic evaluation of the speech quality are functional and produce fully comparable results with the standard evaluation based on the listening test method.

  20. Masking release for words in amplitude-modulated noise as a function of modulation rate and task

    PubMed Central

    Buss, Emily; Whittle, Lisa N.; Grose, John H.; Hall, Joseph W.

    2009-01-01

    For normal-hearing listeners, masked speech recognition can improve with the introduction of masker amplitude modulation. The present experiments tested the hypothesis that this masking release is due in part to an interaction between the temporal distribution of cues necessary to perform the task and the probability of those cues temporally coinciding with masker modulation minima. Stimuli were monosyllabic words masked by speech-shaped noise, and masker modulation was introduced via multiplication with a raised sinusoid of 2.5–40 Hz. Tasks included detection, three-alternative forced-choice identification, and open-set identification. Overall, there was more masking release associated with the closed than the open-set tasks. The best rate of modulation also differed as a function of task; whereas low modulation rates were associated with best performance for the detection and three-alternative identification tasks, performance improved with modulation rate in the open-set task. This task-by-rate interaction was also observed when amplitude-modulated speech was presented in a steady masker, and for low- and high-pass filtered speech presented in modulated noise. These results were interpreted as showing that the optimal rate of amplitude modulation depends on the temporal distribution of speech cues and the information required to perform a particular task. PMID:19603883

  1. Speech: An Opening or a Dead End for Married Women.

    ERIC Educational Resources Information Center

    Forusz, Judith Pulin

    For the American woman, educated and trained in the speech profession, marriage and motherhood induce a shock for which she is unprepared. Society still expects the main responsibility of child rearing to be that of the mother, while the speech profession, which has prepared all students to be teachers and scholars, is uncooperative in providing…

  2. Dopamine Regulation of Human Speech and Bird Song: A Critical Review

    ERIC Educational Resources Information Center

    Simonyan, Kristina; Horwitz, Barry; Jarvis, Erich D.

    2012-01-01

    To understand the neural basis of human speech control, extensive research has been done using a variety of methodologies in a range of experimental models. Nevertheless, several critical questions about learned vocal motor control still remain open. One of them is the mechanism(s) by which neurotransmitters, such as dopamine, modulate speech and…

  3. Do North Carolina Students Have Freedom of Speech? A Review of Campus Speech Codes

    ERIC Educational Resources Information Center

    Robinson, Jenna Ashley

    2010-01-01

    America's colleges and universities are supposed to be strongholds of classically liberal ideals, including the protection of individual rights and openness to debate and inquiry. Too often, this is not the case. Across the country, universities deny students and faculty their fundamental rights to freedom of speech and expression. The report…

  4. Limited mouth opening after primary therapy of head and neck cancer.

    PubMed

    Weber, Clemens; Dommerich, Steffen; Pau, Hans Wilhelm; Kramp, Burkhard

    2010-09-01

    Patients after surgery and radiation/chemoradiation for treatment of head and neck cancer often suffer from oral complications. These problems may be caused by surgery and radiation. Patients complain, for example, of swallowing problems and limited mouth opening (trismus). The maximal interincisal mouth opening (MIO) was measured in patients treated with surgery and radiation/chemoradiation for head and neck cancer at the Department of Otorhinolaryngology at the University of Rostock. These patients also completed a 20-item questionnaire concerning nutritional, sensual, and speech disorders and pain. One hundred one patients (16 female and 85 male) returned the questionnaire and were included in the study. About 50% of the patients had a limited mouth opening (<36 mm); patients with oropharyngeal cancer had a significant higher risk for trismus (p = .024) than patients with other head and neck cancers, especially compared to patients with laryngeal cancer (p = .013). The questionnaire showed that especially patients with oral cancer report about problems with opening the mouth (73%), eating (65%), drinking (73%), xerostomia (92%), speech disorders (68%), and voice (62%). Patients with laryngeal cancer only reported about problems with xerostomia (62%), speech (83%), and voice (90%), similar to patients with pharyngeal cancer. About half of the patients who underwent primary treatment for oral and oropharyngeal cancer developed trismus and reported about problems with opening the mouth, eating, drinking, dry mouth, voice, and speech. Trismus has a negative impact on quality of life and should be a focus in the postoperative management of patients with oral and oropharyngeal cancer, and, if diagnosed, special treatment should be initialized.

  5. Multichannel Speech Enhancement Based on Generalized Gamma Prior Distribution with Its Online Adaptive Estimation

    NASA Astrophysics Data System (ADS)

    Dat, Tran Huy; Takeda, Kazuya; Itakura, Fumitada

    We present a multichannel speech enhancement method based on MAP speech spectral magnitude estimation using a generalized gamma model of speech prior distribution, where the model parameters are adapted from actual noisy speech in a frame-by-frame manner. The utilization of a more general prior distribution with its online adaptive estimation is shown to be effective for speech spectral estimation in noisy environments. Furthermore, the multi-channel information in terms of cross-channel statistics are shown to be useful to better adapt the prior distribution parameters to the actual observation, resulting in better performance of speech enhancement algorithm. We tested the proposed algorithm in an in-car speech database and obtained significant improvements of the speech recognition performance, particularly under non-stationary noise conditions such as music, air-conditioner and open window.

  6. Perception of synthetic speech produced automatically by rule: Intelligibility of eight text-to-speech systems.

    PubMed

    Greene, Beth G; Logan, John S; Pisoni, David B

    1986-03-01

    We present the results of studies designed to measure the segmental intelligibility of eight text-to-speech systems and a natural speech control, using the Modified Rhyme Test (MRT). Results indicated that the voices tested could be grouped into four categories: natural speech, high-quality synthetic speech, moderate-quality synthetic speech, and low-quality synthetic speech. The overall performance of the best synthesis system, DECtalk-Paul, was equivalent to natural speech only in terms of performance on initial consonants. The findings are discussed in terms of recent work investigating the perception of synthetic speech under more severe conditions. Suggestions for future research on improving the quality of synthetic speech are also considered.

  7. Design Foundations for Content-Rich Acoustic Interfaces: Investigating Audemes as Referential Non-Speech Audio Cues

    ERIC Educational Resources Information Center

    Ferati, Mexhid Adem

    2012-01-01

    To access interactive systems, blind and visually impaired users can leverage their auditory senses by using non-speech sounds. The current structure of non-speech sounds, however, is geared toward conveying user interface operations (e.g., opening a file) rather than large theme-based information (e.g., a history passage) and, thus, is ill-suited…

  8. Missed Opportunity? Was Iran s Green Movement an Unconventional Warfare Option?

    DTIC Science & Technology

    2014-12-12

    leadership. Speeches , reports, websites, and foreign documents constituted the majority of usable research. The author assumed accurate translation of...expanding economic influence. The RAND Corporation’s study compiled research from the OpenSource website, scholarly reports, and translated speeches ...constructed from Mir Houssein Mousavi’s speeches . Although difficult to accredit, the manifesto echoed Green Movement leadership ideologies. This work

  9. Effects of Additional Low-Pass-Filtered Speech on Listening Effort for Noise-Band-Vocoded Speech in Quiet and in Noise.

    PubMed

    Pals, Carina; Sarampalis, Anastasios; van Dijk, Mart; Başkent, Deniz

    2018-05-11

    Residual acoustic hearing in electric-acoustic stimulation (EAS) can benefit cochlear implant (CI) users in increased sound quality, speech intelligibility, and improved tolerance to noise. The goal of this study was to investigate whether the low-pass-filtered acoustic speech in simulated EAS can provide the additional benefit of reducing listening effort for the spectrotemporally degraded signal of noise-band-vocoded speech. Listening effort was investigated using a dual-task paradigm as a behavioral measure, and the NASA Task Load indeX as a subjective self-report measure. The primary task of the dual-task paradigm was identification of sentences presented in three experiments at three fixed intelligibility levels: at near-ceiling, 50%, and 79% intelligibility, achieved by manipulating the presence and level of speech-shaped noise in the background. Listening effort for the primary intelligibility task was reflected in the performance on the secondary, visual response time task. Experimental speech processing conditions included monaural or binaural vocoder, with added low-pass-filtered speech (to simulate EAS) or without (to simulate CI). In Experiment 1, in quiet with intelligibility near-ceiling, additional low-pass-filtered speech reduced listening effort compared with binaural vocoder, in line with our expectations, although not compared with monaural vocoder. In Experiments 2 and 3, for speech in noise, added low-pass-filtered speech allowed the desired intelligibility levels to be reached at less favorable speech-to-noise ratios, as expected. It is interesting that this came without the cost of increased listening effort usually associated with poor speech-to-noise ratios; at 50% intelligibility, even a reduction in listening effort on top of the increased tolerance to noise was observed. The NASA Task Load indeX did not capture these differences. The dual-task results provide partial evidence for a potential decrease in listening effort as a result of adding low-frequency acoustic speech to noise-band-vocoded speech. Whether these findings translate to CI users with residual acoustic hearing will need to be addressed in future research because the quality and frequency range of low-frequency acoustic sound available to listeners with hearing loss may differ from our idealized simulations, and additional factors, such as advanced age and varying etiology, may also play a role.This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-No Derivatives License 4.0 (CCBY-NC-ND), where it is permissible to download and share the work provided it is properly cited. The work cannot be changed in any way or used commercially without permission from the journal.

  10. Perception of synthetic speech produced automatically by rule: Intelligibility of eight text-to-speech systems

    PubMed Central

    GREENE, BETH G.; LOGAN, JOHN S.; PISONI, DAVID B.

    2012-01-01

    We present the results of studies designed to measure the segmental intelligibility of eight text-to-speech systems and a natural speech control, using the Modified Rhyme Test (MRT). Results indicated that the voices tested could be grouped into four categories: natural speech, high-quality synthetic speech, moderate-quality synthetic speech, and low-quality synthetic speech. The overall performance of the best synthesis system, DECtalk-Paul, was equivalent to natural speech only in terms of performance on initial consonants. The findings are discussed in terms of recent work investigating the perception of synthetic speech under more severe conditions. Suggestions for future research on improving the quality of synthetic speech are also considered. PMID:23225916

  11. Systematic studies of modified vocalization: the effect of speech rate on speech production measures during metronome-paced speech in persons who stutter.

    PubMed

    Davidow, Jason H

    2014-01-01

    Metronome-paced speech results in the elimination, or substantial reduction, of stuttering moments. The cause of fluency during this fluency-inducing condition is unknown. Several investigations have reported changes in speech pattern characteristics from a control condition to a metronome-paced speech condition, but failure to control speech rate between conditions limits our ability to determine if the changes were necessary for fluency. This study examined the effect of speech rate on several speech production variables during one-syllable-per-beat metronomic speech in order to determine changes that may be important for fluency during this fluency-inducing condition. Thirteen persons who stutter (PWS), aged 18-62 years, completed a series of speaking tasks. Several speech production variables were compared between conditions produced at different metronome beat rates, and between a control condition and a metronome-paced speech condition produced at a rate equal to the control condition. Vowel duration, voice onset time, pressure rise time and phonated intervals were significantly impacted by metronome beat rate. Voice onset time and the percentage of short (30-100 ms) phonated intervals significantly decreased from the control condition to the equivalent rate metronome-paced speech condition. A reduction in the percentage of short phonated intervals may be important for fluency during syllable-based metronome-paced speech for PWS. Future studies should continue examining the necessity of this reduction. In addition, speech rate must be controlled in future fluency-inducing condition studies, including neuroimaging investigations, in order for this research to make a substantial contribution to finding the fluency-inducing mechanism of fluency-inducing conditions. © 2013 Royal College of Speech and Language Therapists.

  12. Automatic speech recognition technology development at ITT Defense Communications Division

    NASA Technical Reports Server (NTRS)

    White, George M.

    1977-01-01

    An assessment of the applications of automatic speech recognition to defense communication systems is presented. Future research efforts include investigations into the following areas: (1) dynamic programming; (2) recognition of speech degraded by noise; (3) speaker independent recognition; (4) large vocabulary recognition; (5) word spotting and continuous speech recognition; and (6) isolated word recognition.

  13. Use of Automated Scoring in Spoken Language Assessments for Test Takers with Speech Impairments. Research Report. ETS RR-17-42

    ERIC Educational Resources Information Center

    Loukina, Anastassia; Buzick, Heather

    2017-01-01

    This study is an evaluation of the performance of automated speech scoring for speakers with documented or suspected speech impairments. Given that the use of automated scoring of open-ended spoken responses is relatively nascent and there is little research to date that includes test takers with disabilities, this small exploratory study focuses…

  14. Separating Contributions of Hearing, Lexical Knowledge, and Speech Production to Speech-Perception Scores in Children with Hearing Impairments.

    ERIC Educational Resources Information Center

    Paatsch, Louise E.; Blamey, Peter J.; Sarant, Julia Z.; Martin, Lois F.A.; Bow, Catherine P.

    2004-01-01

    Open-set word and sentence speech-perception test scores are commonly used as a measure of hearing abilities in children and adults using cochlear implants and/or hearing aids. These tests ore usually presented auditorily with a verbal response. In the case of children, scores are typically lower and more variable than for adults with hearing…

  15. Free Speech and GWOT: Back to the Future?

    DTIC Science & Technology

    2008-02-29

    associated cases from the WWI era) focused on speech as evidence of a substantive crime (there, leaflets were proof that the accused was fomenting the...substantive crime – insurrection within the Army). In Gitlow and Whitney, there was no substantive crime for which speech was the evidence. The...substantive crime was the substance of the speech itself. That this test evaluated the content of the speech itself would later become a major criticism

  16. Systematic Studies of Modified Vocalization: The Effect of Speech Rate on Speech Production Measures During Metronome-Paced Speech in Persons who Stutter

    PubMed Central

    Davidow, Jason H.

    2013-01-01

    Background Metronome-paced speech results in the elimination, or substantial reduction, of stuttering moments. The cause of fluency during this fluency-inducing condition is unknown. Several investigations have reported changes in speech pattern characteristics from a control condition to a metronome-paced speech condition, but failure to control speech rate between conditions limits our ability to determine if the changes were necessary for fluency. Aims This study examined the effect of speech rate on several speech production variables during one-syllable-per-beat metronomic speech, in order to determine changes that may be important for fluency during this fluency-inducing condition. Methods and Procedures Thirteen persons who stutter (PWS), aged 18–62 years, completed a series of speaking tasks. Several speech production variables were compared between conditions produced at different metronome beat rates, and between a control condition and a metronome-paced speech condition produced at a rate equal to the control condition. Outcomes & Results Vowel duration, voice onset time, pressure rise time, and phonated intervals were significantly impacted by metronome beat rate. Voice onset time and the percentage of short (30–100 ms) phonated intervals significantly decreased from the control condition to the equivalent rate metronome-paced speech condition. Conclusions & Implications A reduction in the percentage of short phonated intervals may be important for fluency during syllable-based metronome-paced speech for PWS. Future studies should continue examining the necessity of this reduction. In addition, speech rate must be controlled in future fluency-inducing condition studies, including neuroimaging investigations, in order for this research to make a substantial contribution to finding the fluency-inducing mechanism of fluency-inducing conditions. PMID:24372888

  17. The Silent Battle.

    DTIC Science & Technology

    1982-04-01

    exhibition opened. Lenin stated that freedom of speech and press would be possible only after capitalism was defeated, a classless society was developed, and...countries because property owners (capitalists) controlled the newspapers. but, he said freedom of speech did exist in the Soviet Union because the

  18. Barista: A Framework for Concurrent Speech Processing by USC-SAIL

    PubMed Central

    Can, Doğan; Gibson, James; Vaz, Colin; Georgiou, Panayiotis G.; Narayanan, Shrikanth S.

    2016-01-01

    We present Barista, an open-source framework for concurrent speech processing based on the Kaldi speech recognition toolkit and the libcppa actor library. With Barista, we aim to provide an easy-to-use, extensible framework for constructing highly customizable concurrent (and/or distributed) networks for a variety of speech processing tasks. Each Barista network specifies a flow of data between simple actors, concurrent entities communicating by message passing, modeled after Kaldi tools. Leveraging the fast and reliable concurrency and distribution mechanisms provided by libcppa, Barista lets demanding speech processing tasks, such as real-time speech recognizers and complex training workflows, to be scheduled and executed on parallel (and/or distributed) hardware. Barista is released under the Apache License v2.0. PMID:27610047

  19. Barista: A Framework for Concurrent Speech Processing by USC-SAIL.

    PubMed

    Can, Doğan; Gibson, James; Vaz, Colin; Georgiou, Panayiotis G; Narayanan, Shrikanth S

    2014-05-01

    We present Barista, an open-source framework for concurrent speech processing based on the Kaldi speech recognition toolkit and the libcppa actor library. With Barista, we aim to provide an easy-to-use, extensible framework for constructing highly customizable concurrent (and/or distributed) networks for a variety of speech processing tasks. Each Barista network specifies a flow of data between simple actors, concurrent entities communicating by message passing, modeled after Kaldi tools. Leveraging the fast and reliable concurrency and distribution mechanisms provided by libcppa, Barista lets demanding speech processing tasks, such as real-time speech recognizers and complex training workflows, to be scheduled and executed on parallel (and/or distributed) hardware. Barista is released under the Apache License v2.0.

  20. "The Communication Needs and Rights of Mankind", Group 1 Report of the Futuristic Priorities Division of the Speech Communication Association. "Future Communication Technologies; Hardware and Software"; Group 2 Report.

    ERIC Educational Resources Information Center

    Dance, Frank E. X.; And Others

    This paper reports on the Futuristic Priorities Division members' recommendations and priorities concerning the impact of the future on communication and on the speech communication discipline. The recommendations and priorities are listed for two subgroups: The Communication Needs and Rights of Mankind; and Future Communication Technologies:…

  1. Opinion: Challenging the Future of Education's Discourse

    ERIC Educational Resources Information Center

    Wall, Steven D.

    2010-01-01

    In this article, the author reflects on Education Secretary Arne Duncan's speech at Teachers College, Columbia University, to an audience of pre-service teachers, their professors, graduate students, and others on the future needs of education. The author found that Mr. Duncan's speech did not suffice in addressing the depth of his subject.…

  2. Synthesized Speech Output and Children: A Scoping Review

    ERIC Educational Resources Information Center

    Drager, Kathryn D. R.; Reichle, Joe; Pinkoski, Carrie

    2010-01-01

    Purpose: Many computer-based augmentative and alternative communication systems in use by children have speech output. This article (a) provides a scoping review of the literature addressing the intelligibility and listener comprehension of synthesized speech output with children and (b) discusses future research directions. Method: Studies…

  3. Cortical Responses to Chinese Phonemes in Preschoolers Predict Their Literacy Skills at School Age.

    PubMed

    Hong, Tian; Shuai, Lan; Frost, Stephen J; Landi, Nicole; Pugh, Kenneth R; Shu, Hua

    2018-01-01

    We investigated whether preschoolers with poor phonological awareness (PA) skills had impaired cortical basis for detecting speech feature, and whether speech perception influences future literacy outcomes in preschoolers. We recorded ERP responses to speech in 52 Chinese preschoolers. The results showed that the poor PA group processed speech changes differentially compared to control group in mismatch negativity (MMN) and late discriminative negativity (LDN). Furthermore, speech perception in kindergarten could predict literacy outcomes after literacy acquisition. These suggest that impairment in detecting speech features occurs before formal reading instruction, and that speech perception plays an important role in reading development.

  4. Acoustic and laryngographic measures of the laryngeal reflexes of linguistic prominence and vocal effort in German1

    PubMed Central

    Mooshammer, Christine

    2010-01-01

    This study uses acoustic and physiological measures to compare laryngeal reflexes of global changes in vocal effort to the effects of modulating such aspects of linguistic prominence as sentence accent, induced by focus variation, and word stress. Seven speakers were recorded by using a laryngograph. The laryngographic pulses were preprocessed to normalize time and amplitude. The laryngographic pulse shape was quantified using open and skewness quotients and also by applying a functional version of the principal component analysis. Acoustic measures included the acoustic open quotient and spectral balance in the vowel ∕e∕ during the test syllable. The open quotient and the laryngographic pulse shape indicated a significantly shorter open phase for loud speech than for soft speech. Similar results were found for lexical stress, suggesting that lexical stress and loud speech are produced with a similar voice source mechanism. Stressed syllables were distinguished from unstressed syllables by their open phase and pulse shape, even in the absence of sentence accent. Evidence for laryngeal involvement in signaling focus, independent of fundamental frequency changes, was not as consistent across speakers. Acoustic results on various spectral balance measures were generally much less consistent compared to results from laryngographic data. PMID:20136226

  5. Corporate Speech and the Constitution: The Deregulation of Tobacco Advertising

    PubMed Central

    Gostin, Lawrence O.

    2002-01-01

    In a series of recent cases, the Supreme Court has given businesses powerful new First Amendment rights to advertise hazardous products. Most recently, in Lorillard Tobacco Co v Reilly (121 SCt 2404 [2001]), the court invalidated Massachusetts regulations intended to reduce underage smoking. The future prospects for commercial speech regulation appear dim, but the reasoning in commercial speech cases is supported by only a plurality of the court. A different First Amendment theory should recognize the importance of population health and the low value of corporate speech. In particular, a future court should consider the low informational value of tobacco advertising, the availability of alternative channels of communication, the unlawful practice of targeting minors, and the magnitude of the social harms. PMID:11867306

  6. Corporate speech and the Constitution: the deregulation of tobacco advertising.

    PubMed

    Gostin, Lawrence O

    2002-03-01

    In a series of recent cases, the Supreme Court has given businesses powerful new First Amendment rights to advertise hazardous products. Most recently, in Lorillard Tobacco Co v Reilly (121 SCt 2404 [2001]), the court invalidated Massachusetts regulations intended to reduce underage smoking. The future prospects for commercial speech regulation appear dim, but the reasoning in commercial speech cases is supported by only a plurality of the court. A different First Amendment theory should recognize the importance of population health and the low value of corporate speech. In particular, a future court should consider the low informational value of tobacco advertising, the availability of alternative channels of communication, the unlawful practice of targeting minors, and the magnitude of the social harms.

  7. Proceedings: Cable Broadcasting in the Community. April 30-May 2, 1972.

    ERIC Educational Resources Information Center

    Guelph Univ. (Ontario). Office of Continuing Education.

    The proceedings contain transcripts of speeches, panel discussions, and plenary sessions dealing with various aspects of cable broadcasting. The speeches include: Community Television--Future Potential, John deMercado; Reaction to Dr. deMercado's speech, Diane Abbey Livingston; The Guelph Communications Project, William Foss; An Outline for the…

  8. A Case Study Assessing the Auditory and Speech Development of Four Children Implanted with Cochlear Implants by the Chronological Age of 12 Months

    PubMed Central

    2013-01-01

    Children with severe hearing loss most likely receive the greatest benefit from a cochlear implant (CI) when implanted at less than 2 years of age. Children with a hearing loss may also benefit greater from binaural sensory stimulation. Four children who received their first CI under 12 months of age were included in this study. Effects on auditory development were determined using the German LittlEARS Auditory Questionnaire, closed- and open-set monosyllabic word tests, aided free-field, the Mainzer and Göttinger speech discrimination tests, Monosyllabic-Trochee-Polysyllabic (MTP), and Listening Progress Profile (LiP). Speech production and grammar development were evaluated using a German language speech development test (SETK), reception of grammar test (TROG-D) and active vocabulary test (AWST-R). The data showed that children implanted under 12 months of age reached open-set monosyllabic word discrimination at an age of 24 months. LiP results improved over time, and children recognized 100% of words in the MTP test after 12 months. All children performed as well as or better than their hearing peers in speech production and grammar development. SETK showed that the speech development of these children was in general age appropriate. The data suggests that early hearing loss intervention benefits speech and language development and supports the trend towards early cochlear implantation. Furthermore, the data emphasizes the potential benefits associated with bilateral implantation. PMID:23509653

  9. A case study assessing the auditory and speech development of four children implanted with cochlear implants by the chronological age of 12 months.

    PubMed

    May-Mederake, Birgit; Shehata-Dieler, Wafaa

    2013-01-01

    Children with severe hearing loss most likely receive the greatest benefit from a cochlear implant (CI) when implanted at less than 2 years of age. Children with a hearing loss may also benefit greater from binaural sensory stimulation. Four children who received their first CI under 12 months of age were included in this study. Effects on auditory development were determined using the German LittlEARS Auditory Questionnaire, closed- and open-set monosyllabic word tests, aided free-field, the Mainzer and Göttinger speech discrimination tests, Monosyllabic-Trochee-Polysyllabic (MTP), and Listening Progress Profile (LiP). Speech production and grammar development were evaluated using a German language speech development test (SETK), reception of grammar test (TROG-D) and active vocabulary test (AWST-R). The data showed that children implanted under 12 months of age reached open-set monosyllabic word discrimination at an age of 24 months. LiP results improved over time, and children recognized 100% of words in the MTP test after 12 months. All children performed as well as or better than their hearing peers in speech production and grammar development. SETK showed that the speech development of these children was in general age appropriate. The data suggests that early hearing loss intervention benefits speech and language development and supports the trend towards early cochlear implantation. Furthermore, the data emphasizes the potential benefits associated with bilateral implantation.

  10. Teaching strategies in inclusive classrooms with deaf students.

    PubMed

    Cawthon, S W

    2001-01-01

    The purpose of this study was to investigate teacher speech and educational philosophies in inclusive classrooms with deaf and hearing students. Data were collected from language transcripts, classroom observations, and teacher interviews. Total speech output, Mean Length Utterance, proportion of questions to statements, and proportion of open to closed questions were calculated for each teacher. Teachers directed fewer utterances, on average, to deaf than to hearing students but showed different language patterns on the remaining measures. Inclusive philosophies focused on an individualized approach to teaching, attention to deaf culture, advocacy, smaller class sizes, and an openness to diversity in the classroom. The interpreters' role in the classroom included translating teacher speech, voicing student sign language, mediating communication between deaf students and their peers, and monitoring overall classroom behavior.

  11. Keep Your Windows Open and Mirrors Polished: On Quality Education in a Changing America

    ERIC Educational Resources Information Center

    Katz, Lucinda Lee

    2011-01-01

    Lucinda Lee Katz, head of Marin Country Day School (California), received the 2009 NAIS Diversity Leadership Award. This article presents an edited excerpt of her acceptance speech. In this speech, she outlines what is necessary to move school communities ahead in one's diversity work.

  12. 21 CFR 874.3730 - Laryngeal prosthesis (Taub design).

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... pulmonary air flow to the pharynx in the absence of the larynx, thereby permitting esophageal speech. The device is interposed between openings in the trachea and the esophagus and may be removed and replaced... and over the esophageal mucosa to provide a sound source that is articulated as speech. (b...

  13. Engaged listeners: shared neural processing of powerful political speeches

    PubMed Central

    Häcker, Frank E. K.; Honey, Christopher J.; Hasson, Uri

    2015-01-01

    Powerful speeches can captivate audiences, whereas weaker speeches fail to engage their listeners. What is happening in the brains of a captivated audience? Here, we assess audience-wide functional brain dynamics during listening to speeches of varying rhetorical quality. The speeches were given by German politicians and evaluated as rhetorically powerful or weak. Listening to each of the speeches induced similar neural response time courses, as measured by inter-subject correlation analysis, in widespread brain regions involved in spoken language processing. Crucially, alignment of the time course across listeners was stronger for rhetorically powerful speeches, especially for bilateral regions of the superior temporal gyri and medial prefrontal cortex. Thus, during powerful speeches, listeners as a group are more coupled to each other, suggesting that powerful speeches are more potent in taking control of the listeners’ brain responses. Weaker speeches were processed more heterogeneously, although they still prompted substantially correlated responses. These patterns of coupled neural responses bear resemblance to metaphors of resonance, which are often invoked in discussions of speech impact, and contribute to the literature on auditory attention under natural circumstances. Overall, this approach opens up possibilities for research on the neural mechanisms mediating the reception of entertaining or persuasive messages. PMID:25653012

  14. Five heads are better than one: preliminary results of team-based learning in a communication disorders graduate course.

    PubMed

    Epstein, Baila

    2016-01-01

    Clinical problem-solving is fundamental to the role of the speech-language pathologist in both the diagnostic and treatment processes. The problem-solving often involves collaboration with clients and their families, supervisors, and other professionals. Considering the importance of cooperative problem-solving in the profession, graduate education in speech-language pathology should provide experiences to foster the development of these skills. One evidence-based pedagogical approach that directly targets these abilities is team-based learning (TBL). TBL is a small-group instructional method that focuses on students' in-class application of conceptual knowledge in solving complex problems that they will likely encounter in their future clinical careers. The purpose of this pilot study was to investigate the educational outcomes and students' perceptions of TBL in a communication disorders graduate course on speech and language-based learning disabilities. Nineteen graduate students (mean age = 26 years, SD = 4.93), divided into three groups of five students and one group of four students, who were enrolled in a required graduate course, participated by fulfilling the key components of TBL: individual student preparation; individual and team readiness assurance tests (iRATs and tRATs) that assessed preparedness to apply course content; and application activities that challenged teams to solve complex and authentic clinical problems using course material. Performance on the tRATs was significantly higher than the individual students' scores on the iRATs (p < .001, Cohen's d = 4.08). Students generally reported favourable perceptions of TBL on an end-of-semester questionnaire. Qualitative analysis of responses to open-ended questions organized thematically indicated students' high satisfaction with application activities, discontent with the RATs, and recommendations for increased lecture in the TBL process. The outcomes of this pilot study suggest the effectiveness of TBL as an instructional method that provides student teams with opportunities to apply course content in problem-solving activities followed by immediate feedback. This research also addresses the dearth of empirical information on how graduate programmes in speech-language pathology bridge students' didactic learning and clinical practice. Future studies should examine the utility of this approach in other courses within the field and with more heterogeneous student populations. © 2015 Royal College of Speech and Language Therapists.

  15. Rate and rhythm control strategies for apraxia of speech in nonfluent primary progressive aphasia.

    PubMed

    Beber, Bárbara Costa; Berbert, Monalise Costa Batista; Grawer, Ruth Siqueira; Cardoso, Maria Cristina de Almeida Freitas

    2018-01-01

    The nonfluent/agrammatic variant of primary progressive aphasia is characterized by apraxia of speech and agrammatism. Apraxia of speech limits patients' communication due to slow speaking rate, sound substitutions, articulatory groping, false starts and restarts, segmentation of syllables, and increased difficulty with increasing utterance length. Speech and language therapy is known to benefit individuals with apraxia of speech due to stroke, but little is known about its effects in primary progressive aphasia. This is a case report of a 72-year-old, illiterate housewife, who was diagnosed with nonfluent primary progressive aphasia and received speech and language therapy for apraxia of speech. Rate and rhythm control strategies for apraxia of speech were trained to improve initiation of speech. We discuss the importance of these strategies to alleviate apraxia of speech in this condition and the future perspectives in the area.

  16. A Measure of the Auditory-perceptual Quality of Strain from Electroglottographic Analysis of Continuous Dysphonic Speech: Application to Adductor Spasmodic Dysphonia.

    PubMed

    Somanath, Keerthan; Mau, Ted

    2016-11-01

    (1) To develop an automated algorithm to analyze electroglottographic (EGG) signal in continuous dysphonic speech, and (2) to identify EGG waveform parameters that correlate with the auditory-perceptual quality of strain in the speech of patients with adductor spasmodic dysphonia (ADSD). Software development with application in a prospective controlled study. EGG was recorded from 12 normal speakers and 12 subjects with ADSD reading excerpts from the Rainbow Passage. Data were processed by a new algorithm developed with the specific goal of analyzing continuous dysphonic speech. The contact quotient, pulse width, a new parameter peak skew, and various contact closing slope quotient and contact opening slope quotient measures were extracted. EGG parameters were compared between normal and ADSD speech. Within the ADSD group, intra-subject comparison was also made between perceptually strained syllables and unstrained syllables. The opening slope quotient SO7525 distinguished strained syllables from unstrained syllables in continuous speech within individual subjects with ADSD. The standard deviations, but not the means, of contact quotient, EGGW50, peak skew, and SO7525 were different between normal and ADSD speakers. The strain-stress pattern in continuous speech can be visualized as color gradients based on the variation of EGG parameter values. EGG parameters may provide a within-subject measure of vocal strain and serve as a marker for treatment response. The addition of EGG to multidimensional assessment may lead to improved characterization of the voice disturbance in ADSD. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  17. A measure of the auditory-perceptual quality of strain from electroglottographic analysis of continuous dysphonic speech: Application to adductor spasmodic dysphonia

    PubMed Central

    Somanath, Keerthan; Mau, Ted

    2016-01-01

    Objectives (1) To develop an automated algorithm to analyze electroglottographic (EGG) signal in continuous, dysphonic speech, and (2) to identify EGG waveform parameters that correlate with the auditory-perceptual quality of strain in the speech of patients with adductor spasmodic dysphonia (ADSD). Study Design Software development with application in a prospective controlled study. Methods EGG was recorded from 12 normal speakers and 12 subjects with ADSD reading excerpts from the Rainbow Passage. Data were processed by a new algorithm developed with the specific goal of analyzing continuous dysphonic speech. The contact quotient (CQ), pulse width (EGGW), a new parameter peak skew, and various contact closing slope quotient (SC) and contact opening slope quotient (SO) measures were extracted. EGG parameters were compared between normal and ADSD speech. Within the ADSD group, intra-subject comparison was also made between perceptually strained syllables and unstrained syllables. Results The opening slope quotient SO7525 distinguished strained syllables from unstrained syllables in continuous speech within individual ADSD subjects. The standard deviations, but not the means, of CQ, EGGW50, peak skew, and SO7525 were different between normal and ADSD speakers. The strain-stress pattern in continuous speech can be visualized as color gradients based on the variation of EGG parameter values. Conclusions EGG parameters may provide a within-subject measure of vocal strain and serve as a marker for treatment response. The addition of EGG to multi-dimensional assessment may lead to improved characterization of the voice disturbance in ADSD. PMID:26739857

  18. Visual supports for shared reading with young children: the effect of static overlay design.

    PubMed

    Wood Jackson, Carla; Wahlquist, Jordan; Marquis, Cassandra

    2011-06-01

    This study examined the effects of two types of static overlay design (visual scene display and grid display) on 39 children's use of a speech-generating device during shared storybook reading with an adult. This pilot project included two groups: preschool children with typical communication skills (n = 26) and with complex communication needs (n = 13). All participants engaged in shared reading with two books using each visual layout on a speech-generating device (SGD). The children averaged a greater number of activations when presented with a grid display during introductory exploration and free play. There was a large effect of the static overlay design on the number of silent hits, evidencing more silent hits with visual scene displays. On average, the children demonstrated relatively few spontaneous activations of the speech-generating device while the adult was reading, regardless of overlay design. When responding to questions, children with communication needs appeared to perform better when using visual scene displays, but the effect of display condition on the accuracy of responses to wh-questions was not statistically significant. In response to an open ended question, children with communication disorders demonstrated more frequent activations of the SGD using a grid display than a visual scene. Suggestions for future research as well as potential implications for designing AAC systems for shared reading with young children are discussed.

  19. Cognitive and methodological considerations on the effects of musical expertise on speech segmentation.

    PubMed

    François, Clément; Tillmann, Barbara; Schön, Daniele

    2012-04-01

    Both speech and music are constituted by sequences of sound elements that unfold in time and require listeners to engage cognitive functions such as sequencing, attention, and memory. We recently ran a set of experiments with the aim of testing the effect of musical expertise on a rather high cognitive function: speech segmentation. Here, we will present the main concepts underlying the investigation of speech segmentation as well as its link to music and musical expertise. Interestingly, our results seem to show that musical training and expertise have effects on brain plasticity that may go beyond primary regions. Moreover, to facilitate and improve future research in this domain, we will here describe several delicate methodological precautions that need to be taken into account (e.g., the choice of stimuli, participants, data analyses). Finally, we will give some possible future directions to better understand the impact that music may have on speech processing. © 2012 New York Academy of Sciences.

  20. Open-set speaker identification with diverse-duration speech data

    NASA Astrophysics Data System (ADS)

    Karadaghi, Rawande; Hertlein, Heinz; Ariyaeeinia, Aladdin

    2015-05-01

    The concern in this paper is an important category of applications of open-set speaker identification in criminal investigation, which involves operating with short and varied duration speech. The study presents investigations into the adverse effects of such an operating condition on the accuracy of open-set speaker identification, based on both GMMUBM and i-vector approaches. The experiments are conducted using a protocol developed for the identification task, based on the NIST speaker recognition evaluation corpus of 2008. In order to closely cover the real-world operating conditions in the considered application area, the study includes experiments with various combinations of training and testing data duration. The paper details the characteristics of the experimental investigations conducted and provides a thorough analysis of the results obtained.

  1. Argument, Free Speech, and the Politics of Deliberation in the Composition Classroom.

    ERIC Educational Resources Information Center

    Moneyhun, Clyde

    The classical marketplace metaphor for intellectual exchange forms the ideological basis for the way argument is still taught in composition classrooms, where supposedly students are being prepared to participate as full citizens in an equal democracy. However, such a view of democratic citizenship, free speech, and argument is open to criticism…

  2. Randomized Controlled Trial of Video Self-Modeling Following Speech Restructuring Treatment for Stuttering

    ERIC Educational Resources Information Center

    Cream, Angela; O'Brian, Sue; Jones, Mark; Block, Susan; Harrison, Elisabeth; Lincoln, Michelle; Hewat, Sally; Packman, Ann; Menzies, Ross; Onslow, Mark

    2010-01-01

    Purpose: In this study, the authors investigated the efficacy of video self-modeling (VSM) following speech restructuring treatment to improve the maintenance of treatment effects. Method: The design was an open-plan, parallel-group, randomized controlled trial. Participants were 89 adults and adolescents who undertook intensive speech…

  3. On Being Educated in 1991.

    ERIC Educational Resources Information Center

    Bell, Terrel H.

    This speech was delivered at a conference of student body presidents and college and university presidents. The major purpose of the conference was to help open and maintain clear channels of communication between those who run the colleges and universities, and those who attend them. The major points of the speech were: (1) education is not over…

  4. Investigating Prompt Difficulty in an Automatically Scored Speaking Performance Assessment

    ERIC Educational Resources Information Center

    Cox, Troy L.

    2013-01-01

    Speaking assessments for second language learners have traditionally been expensive to administer because of the cost of rating the speech samples. To reduce the cost, many researchers are investigating the potential of using automatic speech recognition (ASR) as a means to score examinee responses to open-ended prompts. This study examined the…

  5. Open fitting: performance verification of receiver in the ear and receiver in the aid.

    PubMed

    Mondelli, Maria Fernanda Capoani Garcia; Garcia, Tatiana Manfrini; Hashimoto, Fabiana Midori Tokuhara; Rocha, Andressa Vital

    2015-01-01

    To verify the receiver in the ear and receiver in the aid adaptations by measuring in situ the speech perception and users' level of satisfaction. The study was approved by the research ethics committee (Process: 027/2011). Twenty subjects older than 18 years with audiological diagnosis of mild and moderate bilateral descending sensorineural hearing loss were evaluated. The subjects were divided into two groups, where G1 (group 1) was fitted with open-fit hearing aids with the built-in receiver unit (receiver in the ear) and G2 (group 2) was fitted with open-fit hearing aids with RITE. A probe microphone measurement was performed to check the gain and output provided by the amplification and for assessment of speech perception with Hearing in Noise Test with and without hearing aids. After a period of six weeks of use without interruption, the subjects returned for follow-up and answered the Satisfaction with Amplification in Daily Life questionnaire, and were again subjected to Hearing in Noise Test. Both groups presented better test results for speech recognition in the presence of noise. Groups 1 and 2 were satisfied with the use of hearing aids and improved speech recognition in silent and noisy situations with hearing aids. Copyright © 2014 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.

  6. Speech perception and quality of life of open-fit hearing aid users

    PubMed Central

    GARCIA, Tatiana Manfrini; JACOB, Regina Tangerino de Souza; MONDELLI, Maria Fernanda Capoani Garcia

    2016-01-01

    ABSTRACT Objective To relate the performance of individuals with hearing loss at high frequencies in speech perception with the quality of life before and after the fitting of an open-fit hearing aid (HA). Methods The WHOQOL-BREF had been used before the fitting and 90 days after the use of HA. The Hearing in Noise Test (HINT) had been conducted in two phases: (1) at the time of fitting without an HA (situation A) and with an HA (situation B); (2) with an HA 90 days after fitting (situation C). Study Sample Thirty subjects with sensorineural hearing loss at high frequencies. Results By using an analysis of variance and the Tukey’s test comparing the three HINT situations in quiet and noisy environments, an improvement has been observed after the HA fitting. The results of the WHOQOL-BREF have showed an improvement in the quality of life after the HA fitting (paired t-test). The relationship between speech perception and quality of life before the HA fitting indicated a significant relationship between speech recognition in noisy environments and in the domain of social relations after the HA fitting (Pearson’s correlation coefficient). Conclusions The auditory stimulation has improved speech perception and the quality of life of individuals. PMID:27383708

  7. Reaction times of normal listeners to laryngeal, alaryngeal, and synthetic speech.

    PubMed

    Evitts, Paul M; Searl, Jeff

    2006-12-01

    The purpose of this study was to compare listener processing demands when decoding alaryngeal compared to laryngeal speech. Fifty-six listeners were presented with single words produced by 1 proficient speaker from 5 different modes of speech: normal, tracheosophageal (TE), esophageal (ES), electrolaryngeal (EL), and synthetic speech (SS). Cognitive processing load was indexed by listener reaction time (RT). To account for significant durational differences among the modes of speech, an RT ratio was calculated (stimulus duration divided by RT). Results indicated that the cognitive processing load was greater for ES and EL relative to normal speech. TE and normal speech did not differ in terms of RT ratio, suggesting fairly comparable cognitive demands placed on the listener. SS required greater cognitive processing load than normal and alaryngeal speech. The results are discussed relative to alaryngeal speech intelligibility and the role of the listener. Potential clinical applications and directions for future research are also presented.

  8. Intensive speech and language therapy in patients with chronic aphasia after stroke: a randomised, open-label, blinded-endpoint, controlled trial in a health-care setting.

    PubMed

    Breitenstein, Caterina; Grewe, Tanja; Flöel, Agnes; Ziegler, Wolfram; Springer, Luise; Martus, Peter; Huber, Walter; Willmes, Klaus; Ringelstein, E Bernd; Haeusler, Karl Georg; Abel, Stefanie; Glindemann, Ralf; Domahs, Frank; Regenbrecht, Frank; Schlenck, Klaus-Jürgen; Thomas, Marion; Obrig, Hellmuth; de Langen, Ernst; Rocker, Roman; Wigbers, Franziska; Rühmkorf, Christina; Hempen, Indra; List, Jonathan; Baumgaertner, Annette

    2017-04-15

    Treatment guidelines for aphasia recommend intensive speech and language therapy for chronic (≥6 months) aphasia after stroke, but large-scale, class 1 randomised controlled trials on treatment effectiveness are scarce. We aimed to examine whether 3 weeks of intensive speech and language therapy under routine clinical conditions improved verbal communication in daily-life situations in people with chronic aphasia after stroke. In this multicentre, parallel group, superiority, open-label, blinded-endpoint, randomised controlled trial, patients aged 70 years or younger with aphasia after stroke lasting for 6 months or more were recruited from 19 inpatient or outpatient rehabilitation centres in Germany. An external biostatistician used a computer-generated permuted block randomisation method, stratified by treatment centre, to randomly assign participants to either 3 weeks or more of intensive speech and language therapy (≥10 h per week) or 3 weeks deferral of intensive speech and language therapy. The primary endpoint was between-group difference in the change in verbal communication effectiveness in everyday life scenarios (Amsterdam-Nijmegen Everyday Language Test A-scale) from baseline to immediately after 3 weeks of treatment or treatment deferral. All analyses were done using the modified intention-to-treat population (those who received 1 day or more of intensive treatment or treatment deferral). This study is registered with ClinicalTrials.gov, number NCT01540383. We randomly assigned 158 patients between April 1, 2012, and May 31, 2014. The modified intention-to-treat population comprised 156 patients (78 per group). Verbal communication was significantly improved from baseline to after intensive speech and language treatment (mean difference 2·61 points [SD 4·94]; 95% CI 1·49 to 3·72), but not from baseline to after treatment deferral (-0·03 points [4·04]; -0·94 to 0·88; between-group difference Cohen's d 0·58; p=0·0004). Eight patients had adverse events during therapy or treatment deferral (one car accident [in the control group], two common cold [one patient per group], three gastrointestinal or cardiac symptoms [all intervention group], two recurrent stroke [one in intervention group before initiation of treatment, and one before group assignment had occurred]); all were unrelated to study participation. 3 weeks of intensive speech and language therapy significantly enhanced verbal communication in people aged 70 years or younger with chronic aphasia after stroke, providing an effective evidence-based treatment approach in this population. Future studies should examine the minimum treatment intensity required for meaningful treatment effects, and determine whether treatment effects cumulate over repeated intervention periods. German Federal Ministry of Education and Research and the German Society for Aphasia Research and Treatment. Copyright © 2017 Elsevier Ltd. All rights reserved.

  9. Advances in real-time magnetic resonance imaging of the vocal tract for speech science and technology research.

    PubMed

    Toutios, Asterios; Narayanan, Shrikanth S

    2016-01-01

    Real-time magnetic resonance imaging (rtMRI) of the moving vocal tract during running speech production is an important emerging tool for speech production research providing dynamic information of a speaker's upper airway from the entire mid-sagittal plane or any other scan plane of interest. There have been several advances in the development of speech rtMRI and corresponding analysis tools, and their application to domains such as phonetics and phonological theory, articulatory modeling, and speaker characterization. An important recent development has been the open release of a database that includes speech rtMRI data from five male and five female speakers of American English each producing 460 phonetically balanced sentences. The purpose of the present paper is to give an overview and outlook of the advances in rtMRI as a tool for speech research and technology development.

  10. Dopamine regulation of human speech and bird song: A critical review

    PubMed Central

    Simonyan, Kristina; Horwitz, Barry; Jarvis, Erich D.

    2012-01-01

    To understand the neural basis of human speech control, extensive research has been done using a variety of methodologies in a range of experimental models. Nevertheless, several critical questions about learned vocal motor control still remain open. One of them is the mechanism(s) by which neurotransmitters, such as dopamine, modulate speech and song production. In this review, we bring together the two fields of investigations of dopamine action on voice control in humans and songbirds, who share similar behavioral and neural mechanisms for speech and song production. While human studies investigating the role of dopamine in speech control are limited to reports in neurological patients, research on dopaminergic modulation of bird song control has recently expanded our views on how this system might be organized. We discuss the parallels between bird song and human speech from the perspective of dopaminergic control as well as outline important differences between these species. PMID:22284300

  11. Advances in real-time magnetic resonance imaging of the vocal tract for speech science and technology research

    PubMed Central

    TOUTIOS, ASTERIOS; NARAYANAN, SHRIKANTH S.

    2016-01-01

    Real-time magnetic resonance imaging (rtMRI) of the moving vocal tract during running speech production is an important emerging tool for speech production research providing dynamic information of a speaker's upper airway from the entire mid-sagittal plane or any other scan plane of interest. There have been several advances in the development of speech rtMRI and corresponding analysis tools, and their application to domains such as phonetics and phonological theory, articulatory modeling, and speaker characterization. An important recent development has been the open release of a database that includes speech rtMRI data from five male and five female speakers of American English each producing 460 phonetically balanced sentences. The purpose of the present paper is to give an overview and outlook of the advances in rtMRI as a tool for speech research and technology development. PMID:27833745

  12. Engaged listeners: shared neural processing of powerful political speeches.

    PubMed

    Schmälzle, Ralf; Häcker, Frank E K; Honey, Christopher J; Hasson, Uri

    2015-08-01

    Powerful speeches can captivate audiences, whereas weaker speeches fail to engage their listeners. What is happening in the brains of a captivated audience? Here, we assess audience-wide functional brain dynamics during listening to speeches of varying rhetorical quality. The speeches were given by German politicians and evaluated as rhetorically powerful or weak. Listening to each of the speeches induced similar neural response time courses, as measured by inter-subject correlation analysis, in widespread brain regions involved in spoken language processing. Crucially, alignment of the time course across listeners was stronger for rhetorically powerful speeches, especially for bilateral regions of the superior temporal gyri and medial prefrontal cortex. Thus, during powerful speeches, listeners as a group are more coupled to each other, suggesting that powerful speeches are more potent in taking control of the listeners' brain responses. Weaker speeches were processed more heterogeneously, although they still prompted substantially correlated responses. These patterns of coupled neural responses bear resemblance to metaphors of resonance, which are often invoked in discussions of speech impact, and contribute to the literature on auditory attention under natural circumstances. Overall, this approach opens up possibilities for research on the neural mechanisms mediating the reception of entertaining or persuasive messages. © The Author (2015). Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.

  13. Asymmetries for the Visual Expression and Perception of Speech

    ERIC Educational Resources Information Center

    Nicholls, Michael E. R.; Searle, Dara A.

    2006-01-01

    This study explored asymmetries for movement, expression and perception of visual speech. Sixteen dextral models were videoed as they articulated: "bat," "cat," "fat," and "sat." Measurements revealed that the right side of the mouth was opened wider and for a longer period than the left. The asymmetry was accentuated at the beginning and ends of…

  14. Communication and Culture in Ancient India and China.

    ERIC Educational Resources Information Center

    Oliver, Robert T.

    The rhetorical theories and practices of ancient India and China provide the themes of this book. An examination of the relationship between culture and rhetoric, East and West, opens the book. The rhetorical milieu of India, its philosophy, social system, and uses of speech, leads to a probing of the caste system and speech of the Brahmins.…

  15. An Annotated Bibliography of Articles in the "Journal of Speech and Language Pathology-Applied Behavior Analysis"

    ERIC Educational Resources Information Center

    Esch, Barbara E.; Forbes, Heather J.

    2017-01-01

    The open-source "Journal of Speech and Language Pathology-Applied Behavior Analysis" ("JSLP-ABA") was published online from 2006 to 2010. We present an annotated bibliography of 80 articles published in the now-defunct journal with the aim of representing its scholarly content to readers of "The Analysis of Verbal…

  16. Effects of Visual Information on Intelligibility of Open and Closed Class Words in Predictable Sentences Produced by Speakers with Dysarthria

    ERIC Educational Resources Information Center

    Hustad, Katherine C.; Dardis, Caitlin M.; Mccourt, Kelly A.

    2007-01-01

    This study examined the independent and interactive effects of visual information and linguistic class of words on intelligibility of dysarthric speech. Seven speakers with dysarthria participated in the study, along with 224 listeners who transcribed speech samples in audiovisual (AV) or audio-only (AO) listening conditions. Orthographic…

  17. IFLA General Conference, 1990. Official Opening and Plenary Session; Open Forum on IFLA's Core Programmes; Contributed Papers Session. Booklet 0.

    ERIC Educational Resources Information Center

    International Federation of Library Associations, The Hague (Netherlands).

    The 15 papers in this collection include opening remarks, reports on core programs of the International Federation of Library Associations (IFLA), and papers from a contributed papers session. Contents are: (1) "Opening Speech of the IFLA General Conference: Stockholm 1990" (Hans-Peter Geh), in German, with an English translation; (2)…

  18. Anterior open-bite orthodontic treatment in an adult patient: A clinical case report.

    PubMed

    Gracco, Antonio; Siviero, Laura; de Stefani, Alberto; Bruno, Giovanni; Stellini, Edoardo

    2016-06-01

    A 45-year-old woman presented with an anterior open-bite complaining chiefly of her unpleasant smile esthetics and masticatory and speech problems. Treatment included speech therapy initiated immediately after bonding. Lingual spurs were positioned on the mandibular incisors in order to help tongue rehabilitation. During the working phase, temporary anchorage devices (TADs) were used at the mandibular anterior segment to intrude the lower left premolars. A splint was used to ensure retention in the upper and lower arches; an enveloppe linguale nocturne (ELN) was provided. Non-surgical open-bite treatment could offer a valid alternative to orthognanthic surgery when cephalometric evaluation shows no vertical growth pattern; patient compliance is essential to prevent relapse. Copyright © 2016 CEO. Published by Elsevier Masson SAS. All rights reserved.

  19. A common functional neural network for overt production of speech and gesture.

    PubMed

    Marstaller, L; Burianová, H

    2015-01-22

    The perception of co-speech gestures, i.e., hand movements that co-occur with speech, has been investigated by several studies. The results show that the perception of co-speech gestures engages a core set of frontal, temporal, and parietal areas. However, no study has yet investigated the neural processes underlying the production of co-speech gestures. Specifically, it remains an open question whether Broca's area is central to the coordination of speech and gestures as has been suggested previously. The objective of this study was to use functional magnetic resonance imaging to (i) investigate the regional activations underlying overt production of speech, gestures, and co-speech gestures, and (ii) examine functional connectivity with Broca's area. We hypothesized that co-speech gesture production would activate frontal, temporal, and parietal regions that are similar to areas previously found during co-speech gesture perception and that both speech and gesture as well as co-speech gesture production would engage a neural network connected to Broca's area. Whole-brain analysis confirmed our hypothesis and showed that co-speech gesturing did engage brain areas that form part of networks known to subserve language and gesture. Functional connectivity analysis further revealed a functional network connected to Broca's area that is common to speech, gesture, and co-speech gesture production. This network consists of brain areas that play essential roles in motor control, suggesting that the coordination of speech and gesture is mediated by a shared motor control network. Our findings thus lend support to the idea that speech can influence co-speech gesture production on a motoric level. Copyright © 2014 IBRO. Published by Elsevier Ltd. All rights reserved.

  20. Speech in spinocerebellar ataxia.

    PubMed

    Schalling, Ellika; Hartelius, Lena

    2013-12-01

    Spinocerebellar ataxias (SCAs) are a heterogeneous group of autosomal dominant cerebellar ataxias clinically characterized by progressive ataxia, dysarthria and a range of other concomitant neurological symptoms. Only a few studies include detailed characterization of speech symptoms in SCA. Speech symptoms in SCA resemble ataxic dysarthria but symptoms related to phonation may be more prominent. One study to date has shown an association between differences in speech and voice symptoms related to genotype. More studies of speech and voice phenotypes are motivated, to possibly aid in clinical diagnosis. In addition, instrumental speech analysis has been demonstrated to be a reliable measure that may be used to monitor disease progression or therapy outcomes in possible future pharmacological treatments. Intervention by speech and language pathologists should go beyond assessment. Clinical guidelines for management of speech, communication and swallowing need to be developed for individuals with progressive cerebellar ataxia. Copyright © 2013 Elsevier Inc. All rights reserved.

  1. Military applications of automatic speech recognition and future requirements

    NASA Technical Reports Server (NTRS)

    Beek, Bruno; Cupples, Edward J.

    1977-01-01

    An updated summary of the state-of-the-art of automatic speech recognition and its relevance to military applications is provided. A number of potential systems for military applications are under development. These include: (1) digital narrowband communication systems; (2) automatic speech verification; (3) on-line cartographic processing unit; (4) word recognition for militarized tactical data system; and (5) voice recognition and synthesis for aircraft cockpit.

  2. Look Who's Talking: Speech Style and Social Context in Language Input to Infants Are Linked to Concurrent and Future Speech Development

    ERIC Educational Resources Information Center

    Ramírez-Esparza, Nairán; García-Sierra, Adrián; Kuhl, Patricia K.

    2014-01-01

    Language input is necessary for language learning, yet little is known about whether, in natural environments, the speech style and social context of language input to children impacts language development. In the present study we investigated the relationship between language input and language development, examining both the style of parental…

  3. Practical applications of interactive voice technologies: Some accomplishments and prospects

    NASA Technical Reports Server (NTRS)

    Grady, Michael W.; Hicklin, M. B.; Porter, J. E.

    1977-01-01

    A technology assessment of the application of computers and electronics to complex systems is presented. Three existing systems which utilize voice technology (speech recognition and speech generation) are described. Future directions in voice technology are also described.

  4. Addressing the Shortage of Speech-Language Pathologists in School Settings

    ERIC Educational Resources Information Center

    Squires, Katie

    2013-01-01

    There is a shortage of speech-language pathologists (SLPs) in this country. This shortage is due, in part, to the limited number of openings in graduate programs and the increased need for SLPs as their scope of practice widens, the autism rate grows, and the population ages. Schools are feeling this shortage the most. There are several reasons…

  5. Common Schools and Uncommon Conversations: Education, Religious Speech and Public Spaces

    ERIC Educational Resources Information Center

    Strike, Kenneth A.

    2007-01-01

    This paper discusses the role of religious speech in the public square and the common school. It argues for more openness to political theology than many liberals are willing to grant and for an educational strategy of engagement over one of avoidance. The paper argues that the exclusion of religious debate from the public square has dysfunctional…

  6. Bilingual Voicing: A Study of Code-Switching in the Reported Speech of Finnish Immigrants in Estonia

    ERIC Educational Resources Information Center

    Frick, Maria; Riionheimo, Helka

    2013-01-01

    Through a conversation analytic investigation of Finnish-Estonian bilingual (direct) reported speech (i.e., voicing) by Finns who live in Estonia, this study shows how code-switching is used as a double contextualization device. The code-switched voicings are shaped by the on-going interactional situation, serving its needs by opening up a context…

  7. The Contours of Free Expression on Campus: Free Speech, Academic Freedom, and Civility

    ERIC Educational Resources Information Center

    Lawrence, Frederick M.

    2017-01-01

    A tension exists on college and university campuses across America today concerning how to pursue liberal, rational, open learning and, at the same time, celebrate a spirit of academic community--in short, how to exercise free expression and maintain civility. In this article, the author begins with an exploration of the boundaries of free speech,…

  8. Changes of some functional speech disorders after surgical correction of skeletal anterior open bite.

    PubMed

    Knez Ambrožič, Mojca; Hočevar Boltežar, Irena; Ihan Hren, Nataša

    2015-09-01

    Skeletal anterior open bite (AOB) or apertognathism is characterized by the absence of contact of the anterior teeth and affects articulation parameters, chewing, biting and voice quality. The treatment of AOB consists of orthognatic surgical procedures. The aim of this study was to evaluate the effects of treatment on voice quality, articulation and nasality in speech with respect to skeletal changes. The study was prospective; 15 patients with AOB were evaluated before and after surgery. Lateral cephalometric x-ray parameters (facial angle, interincisal distance, Wits appraisal) were measured to determine skeletal changes. Before surgery, nine patients still had articulation disorders despite speech therapy during childhood. The voice quality parameters were determined by acoustic analysis of the vowel sound /a/ (fundamental frequency-F0, jitter, shimmer). Spectral analysis of vowels /a/, /e/, /i/, /o/, /u/ was carried out by determining the mean frequency of the first (F1) and second (F2) formants. Nasality in speech was expressed as the ratio between the nasal and the oral sound energies during speech samples. After surgery, normalizations of facial skeletal parameters were observed in all patients, but no statistically significant changes in articulation and voice quality parameters occurred despite subjective observations of easier articulation. Any deterioration in velopharyngeal insufficiency was absent in all of the patients. In conclusion, the surgical treatment of skeletal AOB does not lead to deterioration in voice, resonance and articulation qualities. Despite surgical correction of the unfavourable skeletal situation of the speech apparatus, the pre-existing articulation disorder cannot improve without professional intervention.

  9. Analysis of Factors Affecting System Performance in the ASpIRE Challenge

    DTIC Science & Technology

    2015-12-13

    performance in the ASpIRE (Automatic Speech recognition In Reverberant Environments) challenge. In particular, overall word error rate (WER) of the solver...systems is analyzed as a function of room, distance between talker and microphone, and microphone type. We also analyze speech activity detection...analysis will inform the design of future challenges and provide insight into the efficacy of current solutions addressing noisy reverberant speech

  10. APEX/SPIN: a free test platform to measure speech intelligibility.

    PubMed

    Francart, Tom; Hofmann, Michael; Vanthornhout, Jonas; Van Deun, Lieselot; van Wieringen, Astrid; Wouters, Jan

    2017-02-01

    Measuring speech intelligibility in quiet and noise is important in clinical practice and research. An easy-to-use free software platform for conducting speech tests is presented, called APEX/SPIN. The APEX/SPIN platform allows the use of any speech material in combination with any noise. A graphical user interface provides control over a large range of parameters, such as number of loudspeakers, signal-to-noise ratio and parameters of the procedure. An easy-to-use graphical interface is provided for calibration and storage of calibration values. To validate the platform, perception of words in quiet and sentences in noise were measured both with APEX/SPIN and with an audiometer and CD player, which is a conventional setup in current clinical practice. Five normal-hearing listeners participated in the experimental evaluation. Speech perception results were similar for the APEX/SPIN platform and conventional procedures. APEX/SPIN is a freely available and open source platform that allows the administration of all kinds of custom speech perception tests and procedures.

  11. The evolution of speech: a comparative review.

    PubMed

    Fitch

    2000-07-01

    The evolution of speech can be studied independently of the evolution of language, with the advantage that most aspects of speech acoustics, physiology and neural control are shared with animals, and thus open to empirical investigation. At least two changes were necessary prerequisites for modern human speech abilities: (1) modification of vocal tract morphology, and (2) development of vocal imitative ability. Despite an extensive literature, attempts to pinpoint the timing of these changes using fossil data have proven inconclusive. However, recent comparative data from nonhuman primates have shed light on the ancestral use of formants (a crucial cue in human speech) to identify individuals and gauge body size. Second, comparative analysis of the diverse vertebrates that have evolved vocal imitation (humans, cetaceans, seals and birds) provides several distinct, testable hypotheses about the adaptive function of vocal mimicry. These developments suggest that, for understanding the evolution of speech, comparative analysis of living species provides a viable alternative to fossil data. However, the neural basis for vocal mimicry and for mimesis in general remains unknown.

  12. Systematic studies of modified vocalization: effects of speech rate and instatement style during metronome stimulation.

    PubMed

    Davidow, Jason H; Bothe, Anne K; Richardson, Jessica D; Andreatta, Richard D

    2010-12-01

    This study introduces a series of systematic investigations intended to clarify the parameters of the fluency-inducing conditions (FICs) in stuttering. Participants included 11 adults, aged 20-63 years, with typical speech-production skills. A repeated measures design was used to examine the relationships between several speech production variables (vowel duration, voice onset time, fundamental frequency, intraoral pressure, pressure rise time, transglottal airflow, and phonated intervals) and speech rate and instatement style during metronome-entrained rhythmic speech. Measures of duration (vowel duration, voice onset time, and pressure rise time) differed across different metronome conditions. When speech rates were matched between the control condition and metronome condition, voice onset time was the only variable that changed. Results confirm that speech rate and instatement style can influence speech production variables during the production of fluency-inducing conditions. Future studies of normally fluent speech and of stuttered speech must control both features and should further explore the importance of voice onset time, which may be influenced by rate during metronome stimulation in a way that the other variables are not.

  13. Datalink in air traffic management: Human factors issues in communications.

    PubMed

    Stedmon, Alex W; Sharples, Sarah; Littlewood, Robert; Cox, Gemma; Patel, Harshada; Wilson, John R

    2007-07-01

    This paper examines issues underpinning the potential move in aviation away from real speech radiotelephony (R/T) communications towards datalink communications involving text and synthetic speech communications. Using a novel air traffic control (ATC) task, two experiments are reported. Experiment 1 compared the use of speech and text while Experiment 2 compared the use of real and synthetic speech communications. Results indicated that generally there were no significant differences between speech and text communications and that either type could be used without any main effects on performance. However, a number of specific differences were observed across the different phases of the scenarios indicating that workload levels may be more varied when speech communications are used. Experiment 2 illustrated that participants placed a greater level of trust in real speech than synthetic speech, and trusted true communications more than false communications (regardless of whether they were real or synthetic voices). The findings are considered in terms of datalink initiatives for future air traffic management, the importance placed on real speech R/T communications, and the need to develop more natural synthetic speech in this application area.

  14. Autonomic Correlates of Speech Versus Nonspeech Tasks in Children and Adults

    PubMed Central

    Arnold, Hayley S.; MacPherson, Megan K.; Smith, Anne

    2015-01-01

    Purpose To assess autonomic arousal associated with speech and nonspeech tasks in school-age children and young adults. Method Measures of autonomic arousal (electrodermal level, electrodermal response amplitude, blood pulse volume, and heart rate) were recorded prior to, during, and after the performance of speech and nonspeech tasks by twenty 7- to 9-year-old children and twenty 18- to 22-year-old adults. Results Across age groups, autonomic arousal was higher for speech tasks compared with nonspeech tasks, based on peak electrodermal response amplitude and blood pulse volume. Children demonstrated greater relative arousal, based on heart rate and blood pulse volume, for nonspeech oral motor tasks than adults but showed similar mean arousal levels for speech tasks as adults. Children demonstrated sex differences in autonomic arousal; specifically, autonomic arousal remained high for school-age boys but not girls in a more complex open-ended narrative task that followed a simple sentence production task. Conclusions Speech tasks elicit greater autonomic arousal than nonspeech tasks, and children demonstrate greater autonomic arousal for nonspeech oral motor tasks than adults. Sex differences in autonomic arousal associated with speech tasks in school-age children are discussed relative to speech-language differences between boys and girls. PMID:24686989

  15. Experimental comparison between speech transmission index, rapid speech transmission index, and speech intelligibility index.

    PubMed

    Larm, Petra; Hongisto, Valtteri

    2006-02-01

    During the acoustical design of, e.g., auditoria or open-plan offices, it is important to know how speech can be perceived in various parts of the room. Different objective methods have been developed to measure and predict speech intelligibility, and these have been extensively used in various spaces. In this study, two such methods were compared, the speech transmission index (STI) and the speech intelligibility index (SII). Also the simplification of the STI, the room acoustics speech transmission index (RASTI), was considered. These quantities are all based on determining an apparent speech-to-noise ratio on selected frequency bands and summing them using a specific weighting. For comparison, some data were needed on the possible differences of these methods resulting from the calculation scheme and also measuring equipment. Their prediction accuracy was also of interest. Measurements were made in a laboratory having adjustable noise level and absorption, and in a real auditorium. It was found that the measurement equipment, especially the selection of the loudspeaker, can greatly affect the accuracy of the results. The prediction accuracy of the RASTI was found acceptable, if the input values for the prediction are accurately known, even though the studied space was not ideally diffuse.

  16. Development of the Russian matrix sentence test.

    PubMed

    Warzybok, Anna; Zokoll, Melanie; Wardenga, Nina; Ozimek, Edward; Boboshko, Maria; Kollmeier, Birger

    2015-01-01

    To develop the Russian matrix sentence test for speech intelligibility measurements in noise. Test development included recordings, optimization of speech material, and evaluation to investigate the equivalency of the test lists and training. For each of the 500 test items, the speech intelligibility function, speech reception threshold (SRT: signal-to-noise ratio, SNR, that provides 50% speech intelligibility), and slope was obtained. The speech material was homogenized by applying level corrections. In evaluation measurements, speech intelligibility was measured at two fixed SNRs to compare list-specific intelligibility functions. To investigate the training effect and establish reference data, speech intelligibility was measured adaptively. Overall, 77 normal-hearing native Russian listeners. The optimization procedure decreased the spread in SRTs across words from 2.8 to 0.6 dB. Evaluation measurements confirmed that the 16 test lists were equivalent, with a mean SRT of -9.5 ± 0.2 dB and a slope of 13.8 ± 1.6%/dB. The reference SRT, -8.8 ± 0.8 dB for the open-set and -9.4 ± 0.8 dB for the closed-set format, increased slightly for noise levels above 75 dB SPL. The Russian matrix sentence test is suitable for accurate and reliable speech intelligibility measurements in noise.

  17. Recording high quality speech during tagged cine-MRI studies using a fiber optic microphone.

    PubMed

    NessAiver, Moriel S; Stone, Maureen; Parthasarathy, Vijay; Kahana, Yuvi; Paritsky, Alexander; Paritsky, Alex

    2006-01-01

    To investigate the feasibility of obtaining high quality speech recordings during cine imaging of tongue movement using a fiber optic microphone. A Complementary Spatial Modulation of Magnetization (C-SPAMM) tagged cine sequence triggered by an electrocardiogram (ECG) simulator was used to image a volunteer while speaking the syllable pairs /a/-/u/, /i/-/u/, and the words "golly" and "Tamil" in sync with the imaging sequence. A noise-canceling, optical microphone was fastened approximately 1-2 inches above the mouth of the volunteer. The microphone was attached via optical fiber to a laptop computer, where the speech was sampled at 44.1 kHz. A reference recording of gradient activity with no speech was subtracted from target recordings. Good quality speech was discernible above the background gradient sound using the fiber optic microphone without reference subtraction. The audio waveform of gradient activity was extremely stable and reproducible. Subtraction of the reference gradient recording further reduced gradient noise by roughly 21 dB, resulting in exceptionally high quality speech waveforms. It is possible to obtain high quality speech recordings using an optical microphone even during exceptionally loud cine imaging sequences. This opens up the possibility of more elaborate MRI studies of speech including spectral analysis of the speech signal in all types of MRI.

  18. Internet images of the speech pathology profession.

    PubMed

    Byrne, Nicole

    2017-06-05

    Objective The Internet provides the general public with information about speech pathology services, including client groups and service delivery models, as well as the professionals providing the services. Although this information assists the general public and other professionals to both access and understand speech pathology services, it also potentially provides information about speech pathology as a prospective career, including the types of people who are speech pathologists (i.e. demographics). The aim of the present study was to collect baseline data on how the speech pathology profession was presented via images on the Internet. Methods A pilot prospective observational study using content analysis methodology was conducted to analyse publicly available Internet images related to the speech pathology profession. The terms 'Speech Pathology' and 'speech pathologist' to represent both the profession and the professional were used, resulting in the identification of 200 images. These images were considered across a range of areas, including who was in the image (e.g. professional, client, significant other), the technology used and the types of intervention. Results The majority of images showed both a client and a professional (i.e. speech pathologist). While the professional was predominantly presented as female, the gender of the client was more evenly distributed. The clients were more likely to be preschool or school aged, however male speech pathologists were presented as providing therapy to selected age groups (i.e. school aged and younger adults). Images were predominantly of individual therapy and the few group images that were presented were all paediatric. Conclusion Current images of speech pathology continue to portray narrow professional demographics and client groups (e.g. paediatrics). Promoting images of wider scope to fully represent the depth and breadth of speech pathology professional practice may assist in attracting a more diverse group of people into the profession in the future. What is known about the topic? To date, research has not considered the promotional profile of allied health professionals on the Internet. There has been a lack of consideration of whether the way in which the professions are promoted may affect clients accessing allied health services or people entering careers. What does this paper add? This paper raises awareness of the lack of promotion of a diverse workforce in speech pathology and considers how this may affect changing the professional demographics in the future. It also provides a starting point for documentation in the form of a baseline for tracking future changes. It allows consideration of the fact that when designing health promotional and educational materials, it is crucial that diversity is displayed in the professional role, the client role and the setting in order to provide information and education to the general public about the health services provided. What are the implications for practitioners? The presentation of narrow demographics of both the professional and client may potentially affect people considering speech pathology as a future career. The appearance of narrow client demographics and diagnosis groups may also deter people from accessing services. For example, if the demonstrated images do not show older people accessing speech pathology services, then this may suggest that services are only for children. The results from the present case example are transferrable to other health professions with similar professional demographic profiles (e.g. occupational therapy). Consideration of the need to display a diverse client profile is relevant to all health and medical services, and demonstrates steps towards inclusiveness and increasing engagement of clients who may be currently less likely to access health services (including people who are Aboriginal or from a culturally and linguistically diverse background).

  19. Measurement of trained speech patterns in stuttering: interjudge and intrajudge agreement of experts by means of modified time-interval analysis.

    PubMed

    Alpermann, Anke; Huber, Walter; Natke, Ulrich; Willmes, Klaus

    2010-09-01

    Improved fluency after stuttering therapy is usually measured by the percentage of stuttered syllables. However, outcome studies rarely evaluate the use of trained speech patterns that speakers use to manage stuttering. This study investigated whether the modified time interval analysis can distinguish between trained speech patterns, fluent speech, and stuttered speech. Seventeen German experts on stuttering judged a speech sample on two occasions. Speakers of the sample were stuttering adults, who were not undergoing therapy, as well as participants in a fluency shaping and a stuttering modification therapy. Results showed satisfactory inter-judge and intra-judge agreement above 80%. Intervals with trained speech patterns were identified as consistently as stuttered and fluent intervals. We discuss limitations of the study, as well as implications of our findings for the development of training for identification of trained speech patterns and future outcome studies. The reader will be able to (a) explain different methods to measure the use of trained speech patterns, (b) evaluate whether German experts are able to discriminate intervals with trained speech patterns reliably from fluent and stuttered intervals and (c) describe how the measurement of trained speech patterns can contribute to outcome studies.

  20. A systematic review of treatment intensity in speech disorders.

    PubMed

    Kaipa, Ramesh; Peterson, Abigail Marie

    2016-12-01

    Treatment intensity (sometimes referred to as "practice amount") has been well-investigated in learning non-speech tasks, but its role in treating speech disorders has not been largely analysed. This study reviewed the literature regarding treatment intensity in speech disorders. A systematic search was conducted in four databases using appropriate search terms. Seven articles from a total of 580 met the inclusion criteria. The speech disorders investigated included speech sound disorders, dysarthria, acquired apraxia of speech and childhood apraxia of speech. All seven studies were evaluated for their methodological quality, research phase and evidence level. Evidence level of reviewed studies ranged from moderate to strong. With regard to the research phase, only one study was considered to be phase III research, which corresponds to the controlled trial phase. The remaining studies were considered to be phase II research, which corresponds to the phase where magnitude of therapeutic effect is assessed. Results suggested that higher treatment intensity was favourable over lower treatment intensity of specific treatment technique(s) for treating childhood apraxia of speech and speech sound (phonological) disorders. Future research should incorporate randomised-controlled designs to establish optimal treatment intensity that is specific to each of the speech disorders.

  1. Opening up to Native Speaker Norms: The Use of /?/ in the Speech of Canadian French Immersion Students

    ERIC Educational Resources Information Center

    Nadasdi, Terry; Vickerman, Alison

    2017-01-01

    Our study examines the extent to which French immersion students use lax /?/ in the same linguistic context as native speakers of Canadian French. Our results show that the lax variant is vanishingly rare in the speech of immersion students and is used by only a small minority of individuals. This is interpreted as a limitation of French immersion…

  2. Learning for Development: Selected Speeches of Sir John Daniel and Colleagues, September 2006-February 2007

    ERIC Educational Resources Information Center

    Commonwealth of Learning, 2007

    2007-01-01

    The third in a series published by the Commonwealth of Learning (COL), this booklet reproduces five addresses and one article from late 2006 and early 2007. This collection of speeches is entitled "Learning for Development" because that is the focus of the work of the COL's work. The addresses presented here were given at the opening and…

  3. The Development of Co-Speech Gesture and Its Semantic Integration with Speech in 6- to 12-Year-Old Children with Autism Spectrum Disorders

    ERIC Educational Resources Information Center

    So, Wing-Chee; Wong, Miranda Kit-Yi; Lui, Ming; Yip, Virginia

    2015-01-01

    Previous work leaves open the question of whether children with autism spectrum disorders aged 6-12?years have delay in producing gestures compared to their typically developing peers. This study examined gestural production among school-aged children in a naturalistic context and how their gestures are semantically related to the accompanying…

  4. How much does language proficiency by non-native listeners influence speech audiometric tests in noise?

    PubMed

    Warzybok, Anna; Brand, Thomas; Wagener, Kirsten C; Kollmeier, Birger

    2015-01-01

    The current study investigates the extent to which the linguistic complexity of three commonly employed speech recognition tests and second language proficiency influence speech recognition thresholds (SRTs) in noise in non-native listeners. SRTs were measured for non-natives and natives using three German speech recognition tests: the digit triplet test (DTT), the Oldenburg sentence test (OLSA), and the Göttingen sentence test (GÖSA). Sixty-four non-native and eight native listeners participated. Non-natives can show native-like SRTs in noise only for the linguistically easy speech material (DTT). Furthermore, the limitation of phonemic-acoustical cues in digit triplets affects speech recognition to the same extent in non-natives and natives. For more complex and less familiar speech materials, non-natives, ranging from basic to advanced proficiency in German, require on average 3-dB better signal-to-noise ratio for the OLSA and 6-dB for the GÖSA to obtain 50% speech recognition compared to native listeners. In clinical audiology, SRT measurements with a closed-set speech test (i.e. DTT for screening or OLSA test for clinical purposes) should be used with non-native listeners rather than open-set speech tests (such as the GÖSA or HINT), especially if a closed-set version in the patient's own native language is available.

  5. Linguistic Recycling and the Open Community.

    ERIC Educational Resources Information Center

    Dasgupta, Probal

    2001-01-01

    Examines linguistic recycling in the context of domestic Esperanto use. Argues that word-meaning recycling reflects the same fundamental principles as sentential recursion, and that a linguistics theoretically sensitive to these principles strengthens practical efforts towards the social goal of an open speech community. (Author/VWL)

  6. The Role of Clinical Experience in Speech-Language Pathologists' Perception of Subphonemic Detail in Children's Speech

    PubMed Central

    Munson, Benjamin; Johnson, Julie M.; Edwards, Jan

    2013-01-01

    Purpose This study examined whether experienced speech-language pathologists differ from inexperienced people in their perception of phonetic detail in children's speech. Method Convenience samples comprising 21 experienced speech-language pathologist and 21 inexperienced listeners participated in a series of tasks in which they made visual-analog scale (VAS) ratings of children's natural productions of target /s/-/θ/, /t/-/k/, and /d/-/ɡ/ in word-initial position. Listeners rated the perception distance between individual productions and ideal productions. Results The experienced listeners' ratings differed from inexperienced listeners' in four ways: they had higher intra-rater reliability, they showed less bias toward a more frequent sound, their ratings were more closely related to the acoustic characteristics of the children's speech, and their responses were related to a different set of predictor variables. Conclusions Results suggest that experience working as a speech-language pathologist leads to better perception of phonetic detail in children's speech. Limitations and future research are discussed. PMID:22230182

  7. Changes in Voice Onset Time and Motor Speech Skills in Children following Motor Speech Therapy: Evidence from /pa/ productions

    PubMed Central

    Yu, Vickie Y.; Kadis, Darren S.; Oh, Anna; Goshulak, Debra; Namasivayam, Aravind; Pukonen, Margit; Kroll, Robert; De Nil, Luc F.; Pang, Elizabeth W.

    2016-01-01

    This study evaluated changes in motor speech control and inter-gestural coordination for children with speech sound disorders (SSD) subsequent to PROMPT (Prompts for Restructuring Oral Muscular Phonetic Targets) intervention. We measured the distribution patterns of voice onset time (VOT) for a voiceless stop (/p/) to examine the changes in inter-gestural coordination. Two standardized tests were used (VMPAC, GFTA-2) to assess the changes in motor speech skills and articulation. Data showed positive changes in patterns of VOT with a lower pattern of variability. All children showed significantly higher scores for VMPAC, but only some children showed higher scores for GFTA-2. Results suggest that the proprioceptive feedback provided through PROMPT had a positive influence on motor speech control and inter-gestural coordination in voicing behavior. This set of VOT data for children with SSD adds to our understanding of the speech characteristics underlying motor speech control. Directions for future studies are discussed. PMID:24446799

  8. Developmental profile of speech-language and communicative functions in an individual with the Preserved Speech Variant of Rett syndrome

    PubMed Central

    Marschik, Peter B.; Vollmann, Ralf; Bartl-Pokorny, Katrin D.; Green, Vanessa A.; van der Meer, Larah; Wolin, Thomas; Einspieler, Christa

    2018-01-01

    Objective We assessed various aspects of speech-language and communicative functions of an individual with the preserved speech variant (PSV) of Rett syndrome (RTT) to describe her developmental profile over a period of 11 years. Methods For this study we incorporated the following data resources and methods to assess speech-language and communicative functions during pre-, peri- and post-regressional development: retrospective video analyses, medical history data, parental checklists and diaries, standardized tests on vocabulary and grammar, spontaneous speech samples, and picture stories to elicit narrative competences. Results Despite achieving speech-language milestones, atypical behaviours were present at all times. We observed a unique developmental speech-language trajectory (including the RTT typical regression) affecting all linguistic and socio-communicative sub-domains in the receptive as well as the expressive modality. Conclusion Future research should take into consideration a potentially considerable discordance between formal and functional language use by interpreting communicative acts on a more cautionary note. PMID:23870013

  9. Developmental profile of speech-language and communicative functions in an individual with the preserved speech variant of Rett syndrome.

    PubMed

    Marschik, Peter B; Vollmann, Ralf; Bartl-Pokorny, Katrin D; Green, Vanessa A; van der Meer, Larah; Wolin, Thomas; Einspieler, Christa

    2014-08-01

    We assessed various aspects of speech-language and communicative functions of an individual with the preserved speech variant of Rett syndrome (RTT) to describe her developmental profile over a period of 11 years. For this study, we incorporated the following data resources and methods to assess speech-language and communicative functions during pre-, peri- and post-regressional development: retrospective video analyses, medical history data, parental checklists and diaries, standardized tests on vocabulary and grammar, spontaneous speech samples and picture stories to elicit narrative competences. Despite achieving speech-language milestones, atypical behaviours were present at all times. We observed a unique developmental speech-language trajectory (including the RTT typical regression) affecting all linguistic and socio-communicative sub-domains in the receptive as well as the expressive modality. Future research should take into consideration a potentially considerable discordance between formal and functional language use by interpreting communicative acts on a more cautionary note.

  10. Amplitude modulation detection with concurrent frequency modulation.

    PubMed

    Nagaraj, Naveen K

    2016-09-01

    Human speech consists of concomitant temporal modulations in amplitude and frequency that are crucial for speech perception. In this study, amplitude modulation (AM) detection thresholds were measured for 550 and 5000 Hz carriers with and without concurrent frequency modulation (FM), at AM rates crucial for speech perception. Results indicate that adding 40 Hz FM interferes with AM detection, more so for 5000 Hz carrier and for frequency deviations exceeding the critical bandwidth of the carrier frequency. These findings suggest that future cochlear implant processors, encoding speech fine-structures may consider limiting the FM to narrow bandwidth and to low frequencies.

  11. Freedom of Speech Newsletter, Volume 4, Number 1, October 1977.

    ERIC Educational Resources Information Center

    Kelley, Michael P., Ed.

    This newsletter features an essay, "Anticipatory Democracy and Citizen Involvement: Strategies for Communication Education in the Future," which discusses strategies for improving citizen involvement and examines ways in which educators can prepare students for constructive citizen involvement. Notes on Speech Communication Association meetings…

  12. Transcranial electric stimulation for the investigation of speech perception and comprehension

    PubMed Central

    Zoefel, Benedikt; Davis, Matthew H.

    2017-01-01

    ABSTRACT Transcranial electric stimulation (tES), comprising transcranial direct current stimulation (tDCS) and transcranial alternating current stimulation (tACS), involves applying weak electrical current to the scalp, which can be used to modulate membrane potentials and thereby modify neural activity. Critically, behavioural or perceptual consequences of this modulation provide evidence for a causal role of neural activity in the stimulated brain region for the observed outcome. We present tES as a tool for the investigation of which neural responses are necessary for successful speech perception and comprehension. We summarise existing studies, along with challenges that need to be overcome, potential solutions, and future directions. We conclude that, although standardised stimulation parameters still need to be established, tES is a promising tool for revealing the neural basis of speech processing. Future research can use this method to explore the causal role of brain regions and neural processes for the perception and comprehension of speech. PMID:28670598

  13. Speech recognition technology: an outlook for human-to-machine interaction.

    PubMed

    Erdel, T; Crooks, S

    2000-01-01

    Speech recognition, as an enabling technology in healthcare-systems computing, is a topic that has been discussed for quite some time, but is just now coming to fruition. Traditionally, speech-recognition software has been constrained by hardware, but improved processors and increased memory capacities are starting to remove some of these limitations. With these barriers removed, companies that create software for the healthcare setting have the opportunity to write more successful applications. Among the criticisms of speech-recognition applications are the high rates of error and steep training curves. However, even in the face of such negative perceptions, there remains significant opportunities for speech recognition to allow healthcare providers and, more specifically, physicians, to work more efficiently and ultimately spend more time with their patients and less time completing necessary documentation. This article will identify opportunities for inclusion of speech-recognition technology in the healthcare setting and examine major categories of speech-recognition software--continuous speech recognition, command and control, and text-to-speech. We will discuss the advantages and disadvantages of each area, the limitations of the software today, and how future trends might affect them.

  14. Auditory midbrain implant: a review.

    PubMed

    Lim, Hubert H; Lenarz, Minoo; Lenarz, Thomas

    2009-09-01

    The auditory midbrain implant (AMI) is a new hearing prosthesis designed for stimulation of the inferior colliculus in deaf patients who cannot sufficiently benefit from cochlear implants. The authors have begun clinical trials in which five patients have been implanted with a single shank AMI array (20 electrodes). The goal of this review is to summarize the development and research that has led to the translation of the AMI from a concept into the first patients. This study presents the rationale and design concept for the AMI as well a summary of the animal safety and feasibility studies that were required for clinical approval. The authors also present the initial surgical, psychophysical, and speech results from the first three implanted patients. Overall, the results have been encouraging in terms of the safety and functionality of the implant. All patients obtain improvements in hearing capabilities on a daily basis. However, performance varies dramatically across patients depending on the implant location within the midbrain with the best performer still not able to achieve open set speech perception without lip-reading cues. Stimulation of the auditory midbrain provides a wide range of level, spectral, and temporal cues, all of which are important for speech understanding, but they do not appear to sufficiently fuse together to enable open set speech perception with the currently used stimulation strategies. Finally, several issues and hypotheses for why current patients obtain limited speech perception along with several feasible solutions for improving AMI implementation are presented.

  15. What Pinnipeds Have to Say about Human Speech, Music, and the Evolution of Rhythm.

    PubMed

    Ravignani, Andrea; Fitch, W Tecumseh; Hanke, Frederike D; Heinrich, Tamara; Hurgitsch, Bettina; Kotz, Sonja A; Scharff, Constance; Stoeger, Angela S; de Boer, Bart

    2016-01-01

    Research on the evolution of human speech and music benefits from hypotheses and data generated in a number of disciplines. The purpose of this article is to illustrate the high relevance of pinniped research for the study of speech, musical rhythm, and their origins, bridging and complementing current research on primates and birds. We briefly discuss speech, vocal learning, and rhythm from an evolutionary and comparative perspective. We review the current state of the art on pinniped communication and behavior relevant to the evolution of human speech and music, showing interesting parallels to hypotheses on rhythmic behavior in early hominids. We suggest future research directions in terms of species to test and empirical data needed.

  16. What Pinnipeds Have to Say about Human Speech, Music, and the Evolution of Rhythm

    PubMed Central

    Ravignani, Andrea; Fitch, W. Tecumseh; Hanke, Frederike D.; Heinrich, Tamara; Hurgitsch, Bettina; Kotz, Sonja A.; Scharff, Constance; Stoeger, Angela S.; de Boer, Bart

    2016-01-01

    Research on the evolution of human speech and music benefits from hypotheses and data generated in a number of disciplines. The purpose of this article is to illustrate the high relevance of pinniped research for the study of speech, musical rhythm, and their origins, bridging and complementing current research on primates and birds. We briefly discuss speech, vocal learning, and rhythm from an evolutionary and comparative perspective. We review the current state of the art on pinniped communication and behavior relevant to the evolution of human speech and music, showing interesting parallels to hypotheses on rhythmic behavior in early hominids. We suggest future research directions in terms of species to test and empirical data needed. PMID:27378843

  17. Inventing Democracy: Future Alternatives for Social Action.

    ERIC Educational Resources Information Center

    Deethardt, John F.

    1983-01-01

    Considers the rational basis for participatory democracy and six ideas designed to embody that conceptual basis. Contends that the mission of speech communication scholars to the civic culture should be an activation of civic competencies and an invention of new places to practice free speech skills. (PD)

  18. Gender differences in binaural speech-evoked auditory brainstem response: are they clinically significant?

    PubMed

    Jalaei, Bahram; Azmi, Mohd Hafiz Afifi Mohd; Zakaria, Mohd Normani

    2018-05-17

    Binaurally evoked auditory evoked potentials have good diagnostic values when testing subjects with central auditory deficits. The literature on speech-evoked auditory brainstem response evoked by binaural stimulation is in fact limited. Gender disparities in speech-evoked auditory brainstem response results have been consistently noted but the magnitude of gender difference has not been reported. The present study aimed to compare the magnitude of gender difference in speech-evoked auditory brainstem response results between monaural and binaural stimulations. A total of 34 healthy Asian adults aged 19-30 years participated in this comparative study. Eighteen of them were females (mean age=23.6±2.3 years) and the remaining sixteen were males (mean age=22.0±2.3 years). For each subject, speech-evoked auditory brainstem response was recorded with the synthesized syllable /da/ presented monaurally and binaurally. While latencies were not affected (p>0.05), the binaural stimulation produced statistically higher speech-evoked auditory brainstem response amplitudes than the monaural stimulation (p<0.05). As revealed by large effect sizes (d>0.80), substantive gender differences were noted in most of speech-evoked auditory brainstem response peaks for both stimulation modes. The magnitude of gender difference between the two stimulation modes revealed some distinct patterns. Based on these clinically significant results, gender-specific normative data are highly recommended when using speech-evoked auditory brainstem response for clinical and future applications. The preliminary normative data provided in the present study can serve as the reference for future studies on this test among Asian adults. Copyright © 2018 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.

  19. Multimodal Speech Capture System for Speech Rehabilitation and Learning.

    PubMed

    Sebkhi, Nordine; Desai, Dhyey; Islam, Mohammad; Lu, Jun; Wilson, Kimberly; Ghovanloo, Maysam

    2017-11-01

    Speech-language pathologists (SLPs) are trained to correct articulation of people diagnosed with motor speech disorders by analyzing articulators' motion and assessing speech outcome while patients speak. To assist SLPs in this task, we are presenting the multimodal speech capture system (MSCS) that records and displays kinematics of key speech articulators, the tongue and lips, along with voice, using unobtrusive methods. Collected speech modalities, tongue motion, lips gestures, and voice are visualized not only in real-time to provide patients with instant feedback but also offline to allow SLPs to perform post-analysis of articulators' motion, particularly the tongue, with its prominent but hardly visible role in articulation. We describe the MSCS hardware and software components, and demonstrate its basic visualization capabilities by a healthy individual repeating the words "Hello World." A proof-of-concept prototype has been successfully developed for this purpose, and will be used in future clinical studies to evaluate its potential impact on accelerating speech rehabilitation by enabling patients to speak naturally. Pattern matching algorithms to be applied to the collected data can provide patients with quantitative and objective feedback on their speech performance, unlike current methods that are mostly subjective, and may vary from one SLP to another.

  20. Dense home-based recordings reveal typical and atypical development of tense/aspect in a child with delayed language development.

    PubMed

    Chin, Iris; Goodwin, Matthew S; Vosoughi, Soroush; Roy, Deb; Naigles, Letitia R

    2018-01-01

    Studies investigating the development of tense/aspect in children with developmental disorders have focused on production frequency and/or relied on short spontaneous speech samples. How children with developmental disorders use future forms/constructions is also unknown. The current study expands this literature by examining frequency, consistency, and productivity of past, present, and future usage, using the Speechome Recorder, which enables collection of dense, longitudinal audio-video recordings of children's speech. Samples were collected longitudinally in a child who was previously diagnosed with autism spectrum disorder, but at the time of the study exhibited only language delay [Audrey], and a typically developing child [Cleo]. While Audrey was comparable to Cleo in frequency and productivity of tense/aspect use, she was atypical in her consistency and production of an unattested future form. Examining additional measures of densely collected speech samples may reveal subtle atypicalities that are missed when relying on only few typical measures of acquisition.

  1. Open-Source Multi-Language Audio Database for Spoken Language Processing Applications

    DTIC Science & Technology

    2012-12-01

    Mandarin, and Russian . Approximately 30 hours of speech were collected for each language. Each passage has been carefully transcribed at the...manual and automatic methods. The Russian passages have not yet been marked at the phonetic level. Another phase of the work was to explore...You Tube. 300 passages were collected in each of three languages—English, Mandarin, and Russian . Approximately 30 hours of speech were

  2. A Spanish matrix sentence test for assessing speech reception thresholds in noise.

    PubMed

    Hochmuth, Sabine; Brand, Thomas; Zokoll, Melanie A; Castro, Franz Zenker; Wardenga, Nina; Kollmeier, Birger

    2012-07-01

    To develop, optimize, and evaluate a new Spanish sentence test in noise. The test comprises a basic matrix of ten names, verbs, numerals, nouns, and adjectives. From this matrix, test lists of ten sentences with an equal syntactical structure can be formed at random, with each list containing the whole speech material. The speech material represents the phoneme distribution of the Spanish language. The test was optimized for measuring speech reception thresholds (SRTs) in noise by adjusting the presentation levels of the individual words. Subsequently, the test was evaluated by independent measurements investigating the training effects, the comparability of test lists, open-set vs. closed-set test format, and performance of listeners of different Spanish varieties. In total, 68 normal-hearing native Spanish-speaking listeners. SRTs measured using an adaptive procedure were 6.2 ± 0.8 dB SNR for the open-set and 7.2 ± 0.7 dB SNR for the closed-set test format. The residual training effect was less than 1 dB after using two double-lists before data collection. No significant differences were found for listeners of different Spanish varieties indicating that the test is applicable to Spanish as well as Latin American listeners. Test lists can be used interchangeably.

  3. The use of electropalatography in the assessment and treatment of acquired motor speech disorders in adults: current knowledge and future directions.

    PubMed

    McAuliffe, Megan J; Ward, Elizabeth C

    2006-01-01

    Electropalatography (EPG) has been employed to measure speech articulation since the mid-1970s. This technique has predominately been used in experimental phonetic research and in the diagnosis and treatment of articulation disorders in children. However, there is a growing body of research employing EPG to diagnose and treat articulatory impairment associated with acquired motor speech disorder (MSD) in adults. The purpose of this paper was to (1) review the findings of studies pertaining to the assessment and treatment of MSDs in adults using EPG, (2) highlight current methodologies employed, and (3) discuss the potential limitations of EPG in the assessment and treatment of MSDs and examine directions for future applied research and treatment studies.

  4. From the analysis of verbal data to the analysis of organizations: organizing as a dialogical process.

    PubMed

    Lorino, Philippe

    2014-12-01

    The analysis of conversational turn-taking and its implications on time (the speaker cannot completely anticipate the future effects of her/his speech) and sociality (the speech is co-produced by the various speakers rather than by the speaking individual) can provide a useful basis to analyze complex organizing processes and collective action: the actor cannot completely anticipate the future effects of her/his acts and the act is co-produced by multiple actors. This translation from verbal to broader classes of interaction stresses the performativity of speeches, the importance of the situation, the role of semiotic mediations to make temporally and spatially distant "ghosts" present in the dialog, and the dissymmetrical relationship between successive conversational turns, due to temporal irreversibility.

  5. [Device for graphic visualization of pressure course in the opening of sphincter of oesophagus mouth in patients after total laryngectomy].

    PubMed

    Musialik, Wojciech; Markowski, Jarosław; Dziubdziela, Włodzimierz; Likus, Wirginia; Swiderek-Kliszewska, Monika; Paluch, Jarosław

    2012-01-01

    Oesophageal speech were described for the first time by Gutzmann in 1909. The diagnostic method based upon checking the pressure causing opening of oesophagus mouth was developed by Seeman and Van Den Bergh. In the initial stage, the method was meant to determine the level, to which the patient can master oesophageal speech. In the course of further investigations, an additional element of diagnostics was noted, which enabled early detection of recurrence of the neoplastic process in the oesopagus mouth area. The material includes 96 patients treated in 2008-2010 at the Department of Otolaryngology, Medical University in Katowice. All patients were classified on the basis of CT of the larynx and the result of histopathological examination to the total laryngectomy. All the patients were subjected to examination by device to visualize the process of opening pressure of the sphincter of the esophagus paragraph. During two years of observation in 28 cases (29%), oesophageal pressure was open his mouth more than 80 mmHg. When performed CT of the neck and histopathology examination recurrence of laryngeal cancer was confirmed in 23 cases (82%). This device is used to determine the best method of voice and speech rehabilitation in patients after total laryngectomy and early diagnosis of laryngeal and/or hypopharynx cancer recurrence. 2012 Polish Otorhinolaryngology - Head and Neck Surgery Society. Published by Elsevier Urban & Partner Sp. z.o.o.

  6. Scaling and universality in the human voice.

    PubMed

    Luque, Jordi; Luque, Bartolo; Lacasa, Lucas

    2015-04-06

    Speech is a distinctive complex feature of human capabilities. In order to understand the physics underlying speech production, in this work, we empirically analyse the statistics of large human speech datasets ranging several languages. We first show that during speech, the energy is unevenly released and power-law distributed, reporting a universal robust Gutenberg-Richter-like law in speech. We further show that such 'earthquakes in speech' show temporal correlations, as the interevent statistics are again power-law distributed. As this feature takes place in the intraphoneme range, we conjecture that the process responsible for this complex phenomenon is not cognitive, but it resides in the physiological (mechanical) mechanisms of speech production. Moreover, we show that these waiting time distributions are scale invariant under a renormalization group transformation, suggesting that the process of speech generation is indeed operating close to a critical point. These results are put in contrast with current paradigms in speech processing, which point towards low dimensional deterministic chaos as the origin of nonlinear traits in speech fluctuations. As these latter fluctuations are indeed the aspects that humanize synthetic speech, these findings may have an impact in future speech synthesis technologies. Results are robust and independent of the communication language or the number of speakers, pointing towards a universal pattern and yet another hint of complexity in human speech. © 2015 The Author(s) Published by the Royal Society. All rights reserved.

  7. [Test set for the evaluation of hearing and speech development after cochlear implantation in children].

    PubMed

    Lamprecht-Dinnesen, A; Sick, U; Sandrieser, P; Illg, A; Lesinski-Schiedat, A; Döring, W H; Müller-Deile, J; Kiefer, J; Matthias, K; Wüst, A; Konradi, E; Riebandt, M; Matulat, P; Von Der Haar-Heise, S; Swart, J; Elixmann, K; Neumann, K; Hildmann, A; Coninx, F; Meyer, V; Gross, M; Kruse, E; Lenarz, T

    2002-10-01

    Since autumn 1998 the multicenter interdisciplinary study group "Test Materials for CI Children" has been compiling a uniform examination tool for evaluation of speech and hearing development after cochlear implantation in childhood. After studying the relevant literature, suitable materials were checked for practical applicability, modified and provided with criteria for execution and break-off. For data acquisition, observation forms for preparation of a PC-version were developed. The evaluation set contains forms for master data with supplements relating to postoperative processes. The hearing tests check supra-threshold hearing with loudness scaling for children, speech comprehension in silence (Mainz and Göttingen Test for Speech Comprehension in Childhood) and phonemic differentiation (Oldenburg Rhyme Test for Children), the central auditory processes of detection, discrimination, identification and recognition (modification of the "Frankfurt Functional Hearing Test for Children") and audiovisual speech perception (Open Paragraph Tracking, Kiel Speech Track Program). The materials for speech and language development comprise phonetics-phonology, lexicon and semantics (LOGO Pronunciation Test), syntax and morphology (analysis of spontaneous speech), language comprehension (Reynell Scales), communication and pragmatics (observation forms). The MAIS and MUSS modified questionnaires are integrated. The evaluation set serves quality assurance and permits factor analysis as well as controls for regularity through the multicenter comparison of long-term developmental trends after cochlear implantation.

  8. Singing in groups for Parkinson's disease (SING-PD): a pilot study of group singing therapy for PD-related voice/speech disorders.

    PubMed

    Shih, Ludy C; Piel, Jordan; Warren, Amanda; Kraics, Lauren; Silver, Althea; Vanderhorst, Veronique; Simon, David K; Tarsy, Daniel

    2012-06-01

    Parkinson's disease related speech and voice impairment have significant impact on quality of life measures. LSVT(®)LOUD voice and speech therapy (Lee Silverman Voice Therapy) has demonstrated scientific efficacy and clinical effectiveness, but musically based voice and speech therapy has been underexplored as a potentially useful method of rehabilitation. We undertook a pilot, open-label study of a group-based singing intervention, consisting of twelve 90-min weekly sessions led by a voice and speech therapist/singing instructor. The primary outcome measure of vocal loudness as measured by sound pressure level (SPL) at 50 cm during connected speech was not significantly different one week after the intervention or at 13 weeks after the intervention. A number of secondary measures reflecting pitch range, phonation time and maximum loudness also were unchanged. Voice related quality of life (VRQOL) and voice handicap index (VHI) also were unchanged. This study suggests that a group singing therapy intervention at this intensity and frequency does not result in significant improvement in objective and subject-rated measures of voice and speech impairment. Copyright © 2012 Elsevier Ltd. All rights reserved.

  9. Selected Speeches and Essays of Mina Pendo Shaughnessy.

    ERIC Educational Resources Information Center

    Shaughnessy, Mina

    1980-01-01

    Presents Mina Shaughnessy's thoughts on why English professors dislike the teaching of writing, what is needed in writing research, the disadvantages of being a writing teacher at an open admissions school, what open admissions policies have revealed about education in general and basic writing instruction in particular, and writing evaluation…

  10. Multilingual Videos for MOOCs and OER

    ERIC Educational Resources Information Center

    Valor Miró, Juan Daniel; Baquero-Arnal, Pau; Civera, Jorge; Turró, Carlos; Juan, Alfons

    2018-01-01

    Massive Open Online Courses (MOOCs) and Open Educational Resources (OER) are rapidly growing, but are not usually offered in multiple languages due to the lack of cost-effective solutions to translate the different objects comprising them and particularly videos. However, current state-of-the-art automatic speech recognition (ASR) and machine…

  11. Research in speech communication.

    PubMed

    Flanagan, J

    1995-10-24

    Advances in digital speech processing are now supporting application and deployment of a variety of speech technologies for human/machine communication. In fact, new businesses are rapidly forming about these technologies. But these capabilities are of little use unless society can afford them. Happily, explosive advances in microelectronics over the past two decades have assured affordable access to this sophistication as well as to the underlying computing technology. The research challenges in speech processing remain in the traditionally identified areas of recognition, synthesis, and coding. These three areas have typically been addressed individually, often with significant isolation among the efforts. But they are all facets of the same fundamental issue--how to represent and quantify the information in the speech signal. This implies deeper understanding of the physics of speech production, the constraints that the conventions of language impose, and the mechanism for information processing in the auditory system. In ongoing research, therefore, we seek more accurate models of speech generation, better computational formulations of language, and realistic perceptual guides for speech processing--along with ways to coalesce the fundamental issues of recognition, synthesis, and coding. Successful solution will yield the long-sought dictation machine, high-quality synthesis from text, and the ultimate in low bit-rate transmission of speech. It will also open the door to language-translating telephony, where the synthetic foreign translation can be in the voice of the originating talker.

  12. The Relationship Between Spectral Modulation Detection and Speech Recognition: Adult Versus Pediatric Cochlear Implant Recipients

    PubMed Central

    Noble, Jack H.; Camarata, Stephen M.; Sunderhaus, Linsey W.; Dwyer, Robert T.; Dawant, Benoit M.; Dietrich, Mary S.; Labadie, Robert F.

    2018-01-01

    Adult cochlear implant (CI) recipients demonstrate a reliable relationship between spectral modulation detection and speech understanding. Prior studies documenting this relationship have focused on postlingually deafened adult CI recipients—leaving an open question regarding the relationship between spectral resolution and speech understanding for adults and children with prelingual onset of deafness. Here, we report CI performance on the measures of speech recognition and spectral modulation detection for 578 CI recipients including 477 postlingual adults, 65 prelingual adults, and 36 prelingual pediatric CI users. The results demonstrated a significant correlation between spectral modulation detection and various measures of speech understanding for 542 adult CI recipients. For 36 pediatric CI recipients, however, there was no significant correlation between spectral modulation detection and speech understanding in quiet or in noise nor was spectral modulation detection significantly correlated with listener age or age at implantation. These findings suggest that pediatric CI recipients might not depend upon spectral resolution for speech understanding in the same manner as adult CI recipients. It is possible that pediatric CI users are making use of different cues, such as those contained within the temporal envelope, to achieve high levels of speech understanding. Further investigation is warranted to investigate the relationship between spectral and temporal resolution and speech recognition to describe the underlying mechanisms driving peripheral auditory processing in pediatric CI users. PMID:29716437

  13. Speech-Language Dissociations, Distractibility, and Childhood Stuttering

    PubMed Central

    Conture, Edward G.; Walden, Tedra A.; Lambert, Warren E.

    2015-01-01

    Purpose This study investigated the relation among speech-language dissociations, attentional distractibility, and childhood stuttering. Method Participants were 82 preschool-age children who stutter (CWS) and 120 who do not stutter (CWNS). Correlation-based statistics (Bates, Appelbaum, Salcedo, Saygin, & Pizzamiglio, 2003) identified dissociations across 5 norm-based speech-language subtests. The Behavioral Style Questionnaire Distractibility subscale measured attentional distractibility. Analyses addressed (a) between-groups differences in the number of children exhibiting speech-language dissociations; (b) between-groups distractibility differences; (c) the relation between distractibility and speech-language dissociations; and (d) whether interactions between distractibility and dissociations predicted the frequency of total, stuttered, and nonstuttered disfluencies. Results More preschool-age CWS exhibited speech-language dissociations compared with CWNS, and more boys exhibited dissociations compared with girls. In addition, male CWS were less distractible than female CWS and female CWNS. For CWS, but not CWNS, less distractibility (i.e., greater attention) was associated with more speech-language dissociations. Last, interactions between distractibility and dissociations did not predict speech disfluencies in CWS or CWNS. Conclusions The present findings suggest that for preschool-age CWS, attentional processes are associated with speech-language dissociations. Future investigations are warranted to better understand the directionality of effect of this association (e.g., inefficient attentional processes → speech-language dissociations vs. inefficient attentional processes ← speech-language dissociations). PMID:26126203

  14. A Prospectus for the Future Development of a Speech Lab: Hypertext Applications.

    ERIC Educational Resources Information Center

    Berube, David M.

    This paper presents a plan for the next generation of speech laboratories which integrates technologies of modern communication in order to improve and modernize the instructional process. The paper first examines the application of intermediate technologies including audio-video recording and playback, computer assisted instruction and testing…

  15. Winston Churchill's "Iron Curtain" Address: Implications for the Present.

    ERIC Educational Resources Information Center

    Bush, George

    1988-01-01

    Evaluates the "Iron Curtain" speech made by Winston Churchill in 1946, discussing its relevance and implications for the present. Examines Churchill's predictions for the future and his assessment of the USSR. Reviews world developments since the speech and proposes foreign policy goals for the next 40 years. (GEA)

  16. Voice Technologies in Libraries: A Look into the Future.

    ERIC Educational Resources Information Center

    Lange, Holley R., Ed.; And Others

    1991-01-01

    Discussion of synthesized speech and voice recognition focuses on a forum that addressed the potential for speech technologies in libraries. Topics discussed by three contributors include possible library applications in technical processing, book receipt, circulation control, and database access; use by disabled and illiterate users; and problems…

  17. The Forces Restructuring Our Future and Outdoor Recreation: Transcription of Keynote Speech.

    ERIC Educational Resources Information Center

    Feather, Frank

    This futurist keynote speech of the National Conference for Outdoor Leaders addresses the social, technological, economic, and political forces that are restructuring the world. The concept of geostrategic thinking has the components of global thinking, futuristic thinking, and seeking opportunities. Important developments include: (1) wealth will…

  18. Creating Future Scientists and Engineers. 2013 Keynote Speech

    ERIC Educational Resources Information Center

    Hicks, Stephen

    2013-01-01

    This article presents a summary of the keynote speech presented at the ITEEA Conference in Columbus, OH, March 4, 2013, by Steven Hicks. Hicks is former Director, Research & Development, Flavor & Fragrance Development Global Capability, for the Procter & Gamble Company. Educated as a chemical engineer, his outside interests include…

  19. Cerebral specialization for speech production in persons with Down syndrome.

    PubMed

    Heath, M; Elliott, D

    1999-09-01

    The study of cerebral specialization in persons with Down syndrome (DS) has revealed an anomalous pattern of organization. Specifically, dichotic listening studies (e.g., Elliott & Weeks, 1993) have suggested a left ear/right hemisphere dominance for speech perception for persons with DS. In the current investigation, the cerebral dominance for speech production was examined using the mouth asymmetry technique. In right-handed, nonhandicapped subjects, mouth asymmetry methodology has shown that during speech, the right side of the mouth opens sooner and to a larger degree then the left side (Graves, Goodglass, & Landis, 1982). The phenomenon of right mouth asymmetry (RMA) is believed to reflect the direct access that the musculature on the right side of the face has to the left hemisphere's speech production systems. This direct access may facilitate the transfer of innervatory patterns to the muscles on the right side of the face. In the present study, the lateralization for speech production was investigated in 10 right-handed participants with DS and 10 nonhandicapped subjects. A RMA at the initiation and end of speech production occurred for subjects in both groups. Surprisingly, the degree of asymmetry between groups did not differ, suggesting that the lateralization of speech production is similar for persons with and persons without DS. These results support the biological dissociation model (Elliott, Weeks, & Elliott, 1987), which holds that persons with DS display a unique dissociation between speech perception (right hemisphere) and speech production (left hemisphere). Copyright 1999 Academic Press.

  20. Private speech of learning disabled and normally achieving children in classroom academic and laboratory contexts.

    PubMed

    Berk, L E; Landau, S

    1993-04-01

    Learning disabled (LD) children are often targets for cognitive-behavioral interventions designed to train them in effective use of a self-directed speech. The purpose of this study was to determine if, indeed, these children display immature private speech in the naturalistic classroom setting. Comparisons were made of the private speech, motor accompaniment to task, and attention of LD and normally achieving classmates during academic seatwork. Setting effects were examined by comparing classroom data with observations during academic seatwork and puzzle solving in the laboratory. Finally, a subgroup of LD children symptomatic of attention-deficit hyperactivity disorder (ADHD) was compared with pure LD and normally achieving controls to determine if the presumed immature private speech is a function of a learning disability or externalizing behavior problems. Results indicated that LD children used more task-relevant private speech than controls, an effect that was especially pronounced for the LD/ADHD subgroup. Use of private speech was setting- and task-specific. Implications for intervention and future research methodology are discussed.

  1. Do age-related word retrieval difficulties appear (or disappear) in connected speech?

    PubMed

    Kavé, Gitit; Goral, Mira

    2017-09-01

    We conducted a comprehensive literature review of studies of word retrieval in connected speech in healthy aging and reviewed relevant aphasia research that could shed light on the aging literature. Four main hypotheses guided the review: (1) Significant retrieval difficulties would lead to reduced output in connected speech. (2) Significant retrieval difficulties would lead to a more limited lexical variety in connected speech. (3) Significant retrieval difficulties would lead to an increase in word substitution errors and in pronoun use as well as to greater dysfluency and hesitation in connected speech. (4) Retrieval difficulties on tests of single-word production would be associated with measures of word retrieval in connected speech. Studies on aging did not confirm these four hypotheses, unlike studies on aphasia that generally did. The review suggests that future research should investigate how context facilitates word production in old age.

  2. Motor-based intervention protocols in treatment of childhood apraxia of speech (CAS)

    PubMed Central

    Maas, Edwin; Gildersleeve-Neumann, Christina; Jakielski, Kathy J.; Stoeckel, Ruth

    2014-01-01

    This paper reviews current trends in treatment for childhood apraxia of speech (CAS), with a particular emphasis on motor-based intervention protocols. The paper first briefly discusses how CAS fits into the typology of speech sound disorders, followed by a discussion of the potential relevance of principles derived from the motor learning literature for CAS treatment. Next, different motor-based treatment protocols are reviewed, along with their evidence base. The paper concludes with a summary and discussion of future research needs. PMID:25313348

  3. Speech recognition systems on the Cell Broadband Engine

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Liu, Y; Jones, H; Vaidya, S

    In this paper we describe our design, implementation, and first results of a prototype connected-phoneme-based speech recognition system on the Cell Broadband Engine{trademark} (Cell/B.E.). Automatic speech recognition decodes speech samples into plain text (other representations are possible) and must process samples at real-time rates. Fortunately, the computational tasks involved in this pipeline are highly data-parallel and can receive significant hardware acceleration from vector-streaming architectures such as the Cell/B.E. Identifying and exploiting these parallelism opportunities is challenging, but also critical to improving system performance. We observed, from our initial performance timings, that a single Cell/B.E. processor can recognize speech from thousandsmore » of simultaneous voice channels in real time--a channel density that is orders-of-magnitude greater than the capacity of existing software speech recognizers based on CPUs (central processing units). This result emphasizes the potential for Cell/B.E.-based speech recognition and will likely lead to the future development of production speech systems using Cell/B.E. clusters.« less

  4. Speech Rate Entrainment in Children and Adults With and Without Autism Spectrum Disorder.

    PubMed

    Wynn, Camille J; Borrie, Stephanie A; Sellers, Tyra P

    2018-05-03

    Conversational entrainment, a phenomenon whereby people modify their behaviors to match their communication partner, has been evidenced as critical to successful conversation. It is plausible that deficits in entrainment contribute to the conversational breakdowns and social difficulties exhibited by people with autism spectrum disorder (ASD). This study examined speech rate entrainment in children and adult populations with and without ASD. Sixty participants including typically developing children, children with ASD, typically developed adults, and adults with ASD participated in a quasi-conversational paradigm with a pseudoconfederate. The confederate's speech rate was digitally manipulated to create slow and fast speech rate conditions. Typically developed adults entrained their speech rate in the quasi-conversational paradigm, using a faster rate during the fast speech rate conditions and a slower rate during the slow speech rate conditions. This entrainment pattern was not evident in adults with ASD or in children populations. Findings suggest that speech rate entrainment is a developmentally acquired skill and offers preliminary evidence of speech rate entrainment deficits in adults with ASD. Impairments in this area may contribute to the conversational breakdowns and social difficulties experienced by this population. Future work is needed to advance this area of inquiry.

  5. The Downside of Greater Lexical Influences: Selectively Poorer Speech Perception in Noise

    PubMed Central

    Xie, Zilong; Tessmer, Rachel; Chandrasekaran, Bharath

    2017-01-01

    Purpose Although lexical information influences phoneme perception, the extent to which reliance on lexical information enhances speech processing in challenging listening environments is unclear. We examined the extent to which individual differences in lexical influences on phonemic processing impact speech processing in maskers containing varying degrees of linguistic information (2-talker babble or pink noise). Method Twenty-nine monolingual English speakers were instructed to ignore the lexical status of spoken syllables (e.g., gift vs. kift) and to only categorize the initial phonemes (/g/ vs. /k/). The same participants then performed speech recognition tasks in the presence of 2-talker babble or pink noise in audio-only and audiovisual conditions. Results Individuals who demonstrated greater lexical influences on phonemic processing experienced greater speech processing difficulties in 2-talker babble than in pink noise. These selective difficulties were present across audio-only and audiovisual conditions. Conclusion Individuals with greater reliance on lexical processes during speech perception exhibit impaired speech recognition in listening conditions in which competing talkers introduce audible linguistic interferences. Future studies should examine the locus of lexical influences/interferences on phonemic processing and speech-in-speech processing. PMID:28586824

  6. Modulations of 'late' event-related brain potentials in humans by dynamic audiovisual speech stimuli.

    PubMed

    Lebib, Riadh; Papo, David; Douiri, Abdel; de Bode, Stella; Gillon Dowens, Margaret; Baudonnière, Pierre-Marie

    2004-11-30

    Lipreading reliably improve speech perception during face-to-face conversation. Within the range of good dubbing, however, adults tolerate some audiovisual (AV) discrepancies and lipreading, then, can give rise to confusion. We used event-related brain potentials (ERPs) to study the perceptual strategies governing the intermodal processing of dynamic and bimodal speech stimuli, either congruently dubbed or not. Electrophysiological analyses revealed that non-coherent audiovisual dubbings modulated in amplitude an endogenous ERP component, the N300, we compared to a 'N400-like effect' reflecting the difficulty to integrate these conflicting pieces of information. This result adds further support for the existence of a cerebral system underlying 'integrative processes' lato sensu. Further studies should take advantage of this 'N400-like effect' with AV speech stimuli to open new perspectives in the domain of psycholinguistics.

  7. The Neurobiological Grounding of Persistent Stuttering: from Structure to Function.

    PubMed

    Neef, Nicole E; Anwander, Alfred; Friederici, Angela D

    2015-09-01

    Neuroimaging and transcranial magnetic stimulation provide insights into the neuronal mechanisms underlying speech disfluencies in chronic persistent stuttering. In the present paper, the goal is not to provide an exhaustive review of existing literature, but rather to highlight robust findings. We, therefore, conducted a meta-analysis of diffusion tensor imaging studies which have recently implicated disrupted white matter connectivity in stuttering. A reduction of fractional anisotropy in persistent stuttering has been reported at several different loci. Our meta-analysis revealed consistent deficits in the left dorsal stream and in the interhemispheric connections between the sensorimotor cortices. In addition, recent fMRI meta-analyses link stuttering to reduced left fronto-parieto-temporal activation while greater fluency is associated with boosted co-activations of right fronto-parieto-temporal areas. However, the physiological foundation of these irregularities is not accessible with MRI. Complementary, transcranial magnetic stimulation (TMS) reveals local excitatory and inhibitory regulation of cortical dynamics. Applied to a speech motor area, TMS revealed reduced speech-planning-related neuronal dynamics at the level of the primary motor cortex in stuttering. Together, this review provides a focused view of the neurobiology of stuttering to date and may guide the rational design of future research. This future needs to account for the perpetual dynamic interactions between auditory, somatosensory, and speech motor circuits that shape fluent speech.

  8. Speech endpoint detection with non-language speech sounds for generic speech processing applications

    NASA Astrophysics Data System (ADS)

    McClain, Matthew; Romanowski, Brian

    2009-05-01

    Non-language speech sounds (NLSS) are sounds produced by humans that do not carry linguistic information. Examples of these sounds are coughs, clicks, breaths, and filled pauses such as "uh" and "um" in English. NLSS are prominent in conversational speech, but can be a significant source of errors in speech processing applications. Traditionally, these sounds are ignored by speech endpoint detection algorithms, where speech regions are identified in the audio signal prior to processing. The ability to filter NLSS as a pre-processing step can significantly enhance the performance of many speech processing applications, such as speaker identification, language identification, and automatic speech recognition. In order to be used in all such applications, NLSS detection must be performed without the use of language models that provide knowledge of the phonology and lexical structure of speech. This is especially relevant to situations where the languages used in the audio are not known apriori. We present the results of preliminary experiments using data from American and British English speakers, in which segments of audio are classified as language speech sounds (LSS) or NLSS using a set of acoustic features designed for language-agnostic NLSS detection and a hidden-Markov model (HMM) to model speech generation. The results of these experiments indicate that the features and model used are capable of detection certain types of NLSS, such as breaths and clicks, while detection of other types of NLSS such as filled pauses will require future research.

  9. Differences in speech and language abilities between children with 22q11.2 deletion syndrome and children with phenotypic features of 22q11.2 deletion syndrome but without microdeletion.

    PubMed

    Rakonjac, Marijana; Cuturilo, Goran; Stevanovic, Milena; Jelicic, Ljiljana; Subotic, Misko; Jovanovic, Ida; Drakulic, Danijela

    2016-08-01

    22q11.2DS is the most common microdeletion syndrome in humans, usually associated with speech and language delay (SLD). Approximately 75% of children with 22q11.2 microdeletion have congenital heart malformations (CHM) which after infant open-heart surgery might lead to SLD. The purpose of this study was to determine whether factors associated with microdeletion contribute to SLD in children with 22q11.2DS. We compared speech and language abilities of two groups of school-aged children: those with 22q11.2 microdeletion (E1) and those with the phenotype resembling 22q11.2DS but without the microdeletion (E2). An age-matched group of typically developing children was also tested. The obtained results revealed that children from group E1 have lower level of speech and language abilities compared to children from group E2 and control group. Additionally, mild to moderate SLD was detected in children from group E2 compared to children from the control group. The obtained results imply that both CHM after infant open-heart surgery and other factors associated with 22q11.2 microdeletion, contribute to SLD in patients with 22q11.2 microdeletion. Based on this, we could postulate that there is/are some potential candidate gene(s), located in the 22q11.2 region, whose function could be important for speech and language development. Copyright © 2016 Elsevier Ltd. All rights reserved.

  10. A Review of Subsequence Time Series Clustering

    PubMed Central

    Teh, Ying Wah

    2014-01-01

    Clustering of subsequence time series remains an open issue in time series clustering. Subsequence time series clustering is used in different fields, such as e-commerce, outlier detection, speech recognition, biological systems, DNA recognition, and text mining. One of the useful fields in the domain of subsequence time series clustering is pattern recognition. To improve this field, a sequence of time series data is used. This paper reviews some definitions and backgrounds related to subsequence time series clustering. The categorization of the literature reviews is divided into three groups: preproof, interproof, and postproof period. Moreover, various state-of-the-art approaches in performing subsequence time series clustering are discussed under each of the following categories. The strengths and weaknesses of the employed methods are evaluated as potential issues for future studies. PMID:25140332

  11. A review of subsequence time series clustering.

    PubMed

    Zolhavarieh, Seyedjamal; Aghabozorgi, Saeed; Teh, Ying Wah

    2014-01-01

    Clustering of subsequence time series remains an open issue in time series clustering. Subsequence time series clustering is used in different fields, such as e-commerce, outlier detection, speech recognition, biological systems, DNA recognition, and text mining. One of the useful fields in the domain of subsequence time series clustering is pattern recognition. To improve this field, a sequence of time series data is used. This paper reviews some definitions and backgrounds related to subsequence time series clustering. The categorization of the literature reviews is divided into three groups: preproof, interproof, and postproof period. Moreover, various state-of-the-art approaches in performing subsequence time series clustering are discussed under each of the following categories. The strengths and weaknesses of the employed methods are evaluated as potential issues for future studies.

  12. Relations among questionnaire and experience sampling measures of inner speech: a smartphone app study

    PubMed Central

    Alderson-Day, Ben; Fernyhough, Charles

    2015-01-01

    Inner speech is often reported to be a common and central part of inner experience, but its true prevalence is unclear. Many questionnaire-based measures appear to lack convergent validity and it has been claimed that they overestimate inner speech in comparison to experience sampling methods (which involve collecting data at random timepoints). The present study compared self-reporting of inner speech collected via a general questionnaire and experience sampling, using data from a custom-made smartphone app (Inner Life). Fifty-one university students completed a generalized self-report measure of inner speech (the Varieties of Inner Speech Questionnaire, VISQ) and responded to at least seven random alerts to report on incidences of inner speech over a 2-week period. Correlations and pairwise comparisons were used to compare generalized endorsements and randomly sampled scores for each VISQ subscale. Significant correlations were observed between general and randomly sampled measures for only two of the four VISQ subscales, and endorsements of inner speech with evaluative or motivational characteristics did not correlate at all across different measures. Endorsement of inner speech items was significantly lower for random sampling compared to generalized self-report, for all VISQ subscales. Exploratory analysis indicated that specific inner speech characteristics were also related to anxiety and future-oriented thinking. PMID:25964773

  13. Current Evaluation of Upper Oesophageal Sphincter Opening in Dysphagia Practice: An International SLT Survey

    ERIC Educational Resources Information Center

    Regan, Julie; Walshe, Margaret; McMahon, Barry P.

    2012-01-01

    Background: The assessment of adequate upper oesophageal sphincter (UOS) opening during swallowing is an integral component of dysphagia evaluation. Aims: To ascertain speech and language therapists' (SLTs) satisfaction with current methods for assessing UOS function in people with dysphagia and to identify challenges encountered by SLTs with UOS…

  14. Automatic speech recognition in air traffic control

    NASA Technical Reports Server (NTRS)

    Karlsson, Joakim

    1990-01-01

    Automatic Speech Recognition (ASR) technology and its application to the Air Traffic Control system are described. The advantages of applying ASR to Air Traffic Control, as well as criteria for choosing a suitable ASR system are presented. Results from previous research and directions for future work at the Flight Transportation Laboratory are outlined.

  15. Surface Electromyography for Speech and Swallowing Systems: Measurement, Analysis, and Interpretation

    ERIC Educational Resources Information Center

    Stepp, Cara E.

    2012-01-01

    Purpose: Applying surface electromyography (sEMG) to the study of voice, speech, and swallowing is becoming increasingly popular. An improved understanding of sEMG and building a consensus as to appropriate methodology will improve future research and clinical applications. Method: An updated review of the theory behind recording sEMG for the…

  16. Imagery, Concept Formation and Creativity--From Past to Future.

    ERIC Educational Resources Information Center

    Silverstein, Ora. N. Asael

    At the center of the conceptual framework there is visual imagery. Man's emotional and mental behavior is built on archetypal symbols that are the source of creative ideas. Native American pictography, in particular, illustrates this in the correlation between gesture speech and verbal speech. The author's research in this area has included a…

  17. "Side-Coaching" the Public Speech: Toward Improvisational Delivery Adjustments "in the Moment"

    ERIC Educational Resources Information Center

    Waisanen, Don J.; Reynolds, Rodney A.

    2008-01-01

    A post-speech evaluation, suggesting improvements for future presentations, is typical in the public speaking course. Moreover, while most public speaking courses emphasize "thinking" about better delivery, little attention is given to what better delivery "feels" like. If the goal is to cultivate a generation of truly extemporaneous speakers,…

  18. Foreign Experience in Mastering Medical Professional Terminology by Foreign Students at Medical Universities

    ERIC Educational Resources Information Center

    Homonyuk, Olena; Avramenko, Natalia

    2017-01-01

    Professional broadcasting of future medical foreign workers in the medical sector is a compulsory component of their professional readiness. The diversity of professional broadcasting functions of foreign students, the skillful use of the entire range of speech functionality, its external expressive attributability of speech; these are the most…

  19. Translating the Future

    ERIC Educational Resources Information Center

    Giroux, Henry A.

    2005-01-01

    This article is a commencement speech delivered by the author on May 26 at Memorial University of Newfoundland, Canada. In his speech, the author focused on affirming public discourse, civic morality, and what it might mean to conduct your lives as engaged citizens attentive to the suffering of others and the fragility of democracy itself. He said…

  20. Melodic Intonation Therapy: Back to Basics for Future Research

    PubMed Central

    Zumbansen, Anna; Peretz, Isabelle; Hébert, Sylvie

    2014-01-01

    We present a critical review of the literature on melodic intonation therapy (MIT), one of the most formalized treatments used by speech-language therapist in Broca’s aphasia. We suggest basic clarifications to enhance the scientific support of this promising treatment. First, therapeutic protocols using singing as a speech facilitation technique are not necessarily MIT. The goal of MIT is to restore propositional speech. The rationale is that patients can learn a new way to speak through singing by using language-capable regions of the right cerebral hemisphere. Eventually, patients are supposed to use this way of speaking permanently but not to sing overtly. We argue that many treatment programs covered in systematic reviews on MIT’s efficacy do not match MIT’s therapeutic goal and rationale. Critically, we identified two main variations of MIT: the French thérapie mélodique et rythmée (TMR) that trains patients to use singing overtly as a facilitation technique in case of speech struggle and palliative versions of MIT that help patients with the most severe expressive deficits produce a limited set of useful, readymade phrases. Second, we distinguish between the immediate effect of singing on speech production and the long-term effect of the entire program on language recovery. Many results in the MIT literature can be explained by this temporal perspective. Finally, we propose that MIT can be viewed as a treatment of apraxia of speech more than aphasia. This issue should be explored in future experimental studies. PMID:24478754

  1. Speech-Like Rhythm in a Voiced and Voiceless Orangutan Call

    PubMed Central

    Lameira, Adriano R.; Hardus, Madeleine E.; Bartlett, Adrian M.; Shumaker, Robert W.; Wich, Serge A.; Menken, Steph B. J.

    2015-01-01

    The evolutionary origins of speech remain obscure. Recently, it was proposed that speech derived from monkey facial signals which exhibit a speech-like rhythm of ∼5 open-close lip cycles per second. In monkeys, these signals may also be vocalized, offering a plausible evolutionary stepping stone towards speech. Three essential predictions remain, however, to be tested to assess this hypothesis' validity; (i) Great apes, our closest relatives, should likewise produce 5Hz-rhythm signals, (ii) speech-like rhythm should involve calls articulatorily similar to consonants and vowels given that speech rhythm is the direct product of stringing together these two basic elements, and (iii) speech-like rhythm should be experience-based. Via cinematic analyses we demonstrate that an ex-entertainment orangutan produces two calls at a speech-like rhythm, coined “clicks” and “faux-speech.” Like voiceless consonants, clicks required no vocal fold action, but did involve independent manoeuvring over lips and tongue. In parallel to vowels, faux-speech showed harmonic and formant modulations, implying vocal fold and supralaryngeal action. This rhythm was several times faster than orangutan chewing rates, as observed in monkeys and humans. Critically, this rhythm was seven-fold faster, and contextually distinct, than any other known rhythmic calls described to date in the largest database of the orangutan repertoire ever assembled. The first two predictions advanced by this study are validated and, based on parsimony and exclusion of potential alternative explanations, initial support is given to the third prediction. Irrespectively of the putative origins of these calls and underlying mechanisms, our findings demonstrate irrevocably that great apes are not respiratorily, articulatorilly, or neurologically constrained for the production of consonant- and vowel-like calls at speech rhythm. Orangutan clicks and faux-speech confirm the importance of rhythmic speech antecedents within the primate lineage, and highlight potential articulatory homologies between great ape calls and human consonants and vowels. PMID:25569211

  2. Research in speech communication.

    PubMed Central

    Flanagan, J

    1995-01-01

    Advances in digital speech processing are now supporting application and deployment of a variety of speech technologies for human/machine communication. In fact, new businesses are rapidly forming about these technologies. But these capabilities are of little use unless society can afford them. Happily, explosive advances in microelectronics over the past two decades have assured affordable access to this sophistication as well as to the underlying computing technology. The research challenges in speech processing remain in the traditionally identified areas of recognition, synthesis, and coding. These three areas have typically been addressed individually, often with significant isolation among the efforts. But they are all facets of the same fundamental issue--how to represent and quantify the information in the speech signal. This implies deeper understanding of the physics of speech production, the constraints that the conventions of language impose, and the mechanism for information processing in the auditory system. In ongoing research, therefore, we seek more accurate models of speech generation, better computational formulations of language, and realistic perceptual guides for speech processing--along with ways to coalesce the fundamental issues of recognition, synthesis, and coding. Successful solution will yield the long-sought dictation machine, high-quality synthesis from text, and the ultimate in low bit-rate transmission of speech. It will also open the door to language-translating telephony, where the synthetic foreign translation can be in the voice of the originating talker. Images Fig. 1 Fig. 2 Fig. 5 Fig. 8 Fig. 11 Fig. 12 Fig. 13 PMID:7479806

  3. Internet video telephony allows speech reading by deaf individuals and improves speech perception by cochlear implant users.

    PubMed

    Mantokoudis, Georgios; Dähler, Claudia; Dubach, Patrick; Kompis, Martin; Caversaccio, Marco D; Senn, Pascal

    2013-01-01

    To analyze speech reading through Internet video calls by profoundly hearing-impaired individuals and cochlear implant (CI) users. Speech reading skills of 14 deaf adults and 21 CI users were assessed using the Hochmair Schulz Moser (HSM) sentence test. We presented video simulations using different video resolutions (1280 × 720, 640 × 480, 320 × 240, 160 × 120 px), frame rates (30, 20, 10, 7, 5 frames per second (fps)), speech velocities (three different speakers), webcameras (Logitech Pro9000, C600 and C500) and image/sound delays (0-500 ms). All video simulations were presented with and without sound and in two screen sizes. Additionally, scores for live Skype™ video connection and live face-to-face communication were assessed. Higher frame rate (>7 fps), higher camera resolution (>640 × 480 px) and shorter picture/sound delay (<100 ms) were associated with increased speech perception scores. Scores were strongly dependent on the speaker but were not influenced by physical properties of the camera optics or the full screen mode. There is a significant median gain of +8.5%pts (p = 0.009) in speech perception for all 21 CI-users if visual cues are additionally shown. CI users with poor open set speech perception scores (n = 11) showed the greatest benefit under combined audio-visual presentation (median speech perception +11.8%pts, p = 0.032). Webcameras have the potential to improve telecommunication of hearing-impaired individuals.

  4. Linguistic Processing of Accented Speech Across the Lifespan

    PubMed Central

    Cristia, Alejandrina; Seidl, Amanda; Vaughn, Charlotte; Schmale, Rachel; Bradlow, Ann; Floccia, Caroline

    2012-01-01

    In most of the world, people have regular exposure to multiple accents. Therefore, learning to quickly process accented speech is a prerequisite to successful communication. In this paper, we examine work on the perception of accented speech across the lifespan, from early infancy to late adulthood. Unfamiliar accents initially impair linguistic processing by infants, children, younger adults, and older adults, but listeners of all ages come to adapt to accented speech. Emergent research also goes beyond these perceptual abilities, by assessing links with production and the relative contributions of linguistic knowledge and general cognitive skills. We conclude by underlining points of convergence across ages, and the gaps left to face in future work. PMID:23162513

  5. Teachers' perceptions of students with speech sound disorders: a quantitative and qualitative analysis.

    PubMed

    Overby, Megan; Carrell, Thomas; Bernthal, John

    2007-10-01

    This study examined 2nd-grade teachers' perceptions of the academic, social, and behavioral competence of students with speech sound disorders (SSDs). Forty-eight 2nd-grade teachers listened to 2 groups of sentences differing by intelligibility and pitch but spoken by a single 2nd grader. For each sentence group, teachers rated the speaker's academic, social, and behavioral competence using an adapted version of the Teacher Rating Scale of the Self-Perception Profile for Children (S. Harter, 1985) and completed 3 open-ended questions. The matched-guise design controlled for confounding speaker and stimuli variables that were inherent in prior studies. Statistically significant differences in teachers' expectations of children's academic, social, and behavioral performances were found between moderately intelligible and normal intelligibility speech. Teachers associated moderately intelligible low-pitched speech with more behavior problems than moderately intelligible high-pitched speech or either pitch with normal intelligibility. One third of the teachers reported that they could not accurately predict a child's school performance based on the child's speech skills, one third of the teachers causally related school difficulty to SSD, and one third of the teachers made no comment. Intelligibility and speaker pitch appear to be speech variables that influence teachers' perceptions of children's school performance.

  6. DIMED 86: Discurso dos Media e Ensino a Distancia = Discours des Media et Enseignement a Distance = Media Speech and Distance Teaching. Papers Presented at a Seminar (21st, Algarve, Portugal, March 10-15, 1986).

    ERIC Educational Resources Information Center

    Coelho, Maria Eduarda Leal

    Presentations at this seminar on distance education focused on the different types of speech in multimedia presentations which contribute to the elaboration (simulated or substituted) of a situation involving different relationships. In addition to opening and closing remarks by Marcel de Greve and a final report by the Scientific Committee of…

  7. Open Microphone Speech Understanding: Correct Discrimination Of In Domain Speech

    NASA Technical Reports Server (NTRS)

    Hieronymus, James; Aist, Greg; Dowding, John

    2006-01-01

    An ideal spoken dialogue system listens continually and determines which utterances were spoken to it, understands them and responds appropriately while ignoring the rest This paper outlines a simple method for achieving this goal which involves trading a slightly higher false rejection rate of in domain utterances for a higher correct rejection rate of Out of Domain (OOD) utterances. The system recognizes semantic entities specified by a unification grammar which is specialized by Explanation Based Learning (EBL). so that it only uses rules which are seen in the training data. The resulting grammar has probabilities assigned to each construct so that overgeneralizations are not a problem. The resulting system only recognizes utterances which reduce to a valid logical form which has meaning for the system and rejects the rest. A class N-gram grammar has been trained on the same training data. This system gives good recognition performance and offers good Out of Domain discrimination when combined with the semantic analysis. The resulting systems were tested on a Space Station Robot Dialogue Speech Database and a subset of the OGI conversational speech database. Both systems run in real time on a PC laptop and the present performance allows continuous listening with an acceptably low false acceptance rate. This type of open microphone system has been used in the Clarissa procedure reading and navigation spoken dialogue system which is being tested on the International Space Station.

  8. Clear Speech Modifications in Children Aged 6-10

    NASA Astrophysics Data System (ADS)

    Taylor, Griffin Lijding

    Modifications to speech production made by adult talkers in response to instructions to speak clearly have been well documented in the literature. Targeting adult populations has been motivated by efforts to improve speech production for the benefit of the communication partners, however, many adults also have communication partners who are children. Surprisingly, there is limited literature on whether children can change their speech production when cued to speak clearly. Pettinato, Tuomainen, Granlund, and Hazan (2016) showed that by age 12, children exhibited enlarged vowel space areas and reduced articulation rate when prompted to speak clearly, but did not produce any other adult-like clear speech modifications in connected speech. Moreover, Syrett and Kawahara (2013) suggested that preschoolers produced longer and more intense vowels when prompted to speak clearly at the word level. These findings contrasted with adult talkers who show significant temporal and spectral differences between speech produced in control and clear speech conditions. Therefore, it was the purpose of this study to analyze changes in temporal and spectral characteristics of speech production that children aged 6-10 made in these experimental conditions. It is important to elucidate the clear speech profile of this population to better understand which adult-like clear speech modifications they make spontaneously and which modifications are still developing. Understanding these baselines will advance future studies that measure the impact of more explicit instructions and children's abilities to better accommodate their interlocutors, which is a critical component of children's pragmatic and speech-motor development.

  9. The International Space Station: A Pathway to the Future

    NASA Technical Reports Server (NTRS)

    Kitmacher, Gary H.; Gerstenmaier, William H.; Bartoe, John-David F.; Mustachio, Nicholas

    2004-01-01

    Nearly six years after the launch of the first International Space Station element, and four years after its initial occupation, the United States and our 16 international partners have made great strides in operating this impressive Earth orbiting research facility. This past year we have done so in the face of the adversity of operating without the benefit of the Space Shuttle. In his January 14, 2004, speech announcing a new vision for America's space program, President Bush affirmed the United States' commitment to completing construction of the International Space Station by 2010. The President also stated that we would focus our future research aboard the Station on the longterm effects of space travel on human biology. This research will help enable human crews to venture through the vast voids of space for months at a time. In addition, ISS affords a unique opportunity to serve as an engineering test bed for hardware and operations critical to the exploration tasks. NASA looks forward to working with our partners on International Space Station research that will help open up new pathways for future exploration and discovery beyond low Earth orbit. This paper provides an overview of the International Space Station Program focusing on a review of the events of the past year, as well as plans for next year and the future.

  10. Musical expertise and second language learning.

    PubMed

    Chobert, Julie; Besson, Mireille

    2013-06-06

    Increasing evidence suggests that musical expertise influences brain organization and brain functions. Moreover, results at the behavioral and neurophysiological levels reveal that musical expertise positively influences several aspects of speech processing, from auditory perception to speech production. In this review, we focus on the main results of the literature that led to the idea that musical expertise may benefit second language acquisition. We discuss several interpretations that may account for the influence of musical expertise on speech processing in native and foreign languages, and we propose new directions for future research.

  11. Musical Expertise and Second Language Learning

    PubMed Central

    Chobert, Julie; Besson, Mireille

    2013-01-01

    Increasing evidence suggests that musical expertise influences brain organization and brain functions. Moreover, results at the behavioral and neurophysiological levels reveal that musical expertise positively influences several aspects of speech processing, from auditory perception to speech production. In this review, we focus on the main results of the literature that led to the idea that musical expertise may benefit second language acquisition. We discuss several interpretations that may account for the influence of musical expertise on speech processing in native and foreign languages, and we propose new directions for future research. PMID:24961431

  12. Management of non-progressive dysarthria: practice patterns of speech and language therapists in the Republic of Ireland.

    PubMed

    Conway, Aifric; Walshe, Margaret

    2015-01-01

    Dysarthria is a commonly acquired speech disorder. Rising numbers of people surviving stroke and traumatic brain injury (TBI) mean the numbers of people with non-progressive dysarthria are likely to increase, with increased challenges for speech and language therapists (SLTs), service providers and key stakeholders. The evidence base for assessment and intervention approaches with this population remains limited with clinical guidelines relying largely on clinical experience, expert opinion and limited research. Furthermore, there is currently little evidence on the practice behaviours of SLTs available. To investigate whether SLTs in the Republic of Ireland (ROI) vary in how they assess and manage adults with non-progressive dysarthria; to explore SLTs' use of the theoretical principles that influence therapeutic approaches; to identify challenges perceived by SLTs when working with adults with non-progressive dysarthria; and to determine SLTs' perceptions of further training needs. A 33-item survey questionnaire was devised and disseminated electronically via SurveyMonkey to SLTs working with non-progressive dysarthria in the ROI. SLTs were identified through e-mail lists for special-interest groups, SLT manager groups and general SLT mailing lists. A reminder e-mail was sent to all SLTs 3 weeks later following the initial e-mail containing the survey link. The survey remained open for 6 weeks. Questionnaire responses were analysed using descriptive statistics. Qualitative comments to open-ended questions were analysed through thematic analysis. Eighty SLTs responded to the survey. Sixty-seven of these completed the survey in full. SLTs provided both quantitative and qualitative data regarding their assessment and management practices in this area. Practice varied depending on the context of the SLT service, experience of SLTs and the resources available to them. Not all SLTs used principles such as motor programming or neural plasticity to direct clinical work and some requested further direction in this area. SLTs perceived that the key challenges associated with working in this area were the compliance, insight and motivation of adults with dysarthria. The use of specific treatment programmes varies amongst SLTs. A lack of resources is reported to restrict practice in both assessment and management. Ongoing research into the effectiveness of SLT interventions with adults with non-progressive dysarthria is required to guide clinical decision-making. SLTs identified further training needs which may provide direction for the development of professional training courses in the future. © 2015 Royal College of Speech and Language Therapists.

  13. Children's views of communication and speech-language pathology.

    PubMed

    Merrick, Rosalind; Roulstone, Sue

    2011-08-01

    Children have the right to express their views and influence decisions in matters that affect them. Yet decisions regarding speech-language pathology are often made on their behalf, and research into the perspectives of children who receive speech-language pathology intervention is currently limited. This paper reports a qualitative study which explored experiences of communication and of speech-language pathology from the perspectives of children with speech, language, and communication needs (SLCN). The aim was to explore their perspectives of communication, communication impairment, and assistance. Eleven school-children participated in the study, aged between 7-10 years. They were recruited through a speech-language pathology service in south west England, to include a range of ages and severity of difficulties. The study used open-ended interviews within which non-verbal activities such as drawing, taking photographs, and compiling a scrapbook were used to create a context for supported conversations. Findings were analysed according to the principles of grounded theory. Three ways of talking about communication emerged. These were in terms of impairment, learning, and behaviour. Findings offer insight into dialogue between children with SLCN and adults; the way communication is talked about has implications for children's view of themselves, their skills, and their participation.

  14. Voice recognition through phonetic features with Punjabi utterances

    NASA Astrophysics Data System (ADS)

    Kaur, Jasdeep; Juglan, K. C.; Sharma, Vishal; Upadhyay, R. K.

    2017-07-01

    This paper deals with perception and disorders of speech in view of Punjabi language. Visualizing the importance of voice identification, various parameters of speaker identification has been studied. The speech material was recorded with a tape recorder in their normal and disguised mode of utterances. Out of the recorded speech materials, the utterances free from noise, etc were selected for their auditory and acoustic spectrographic analysis. The comparison of normal and disguised speech of seven subjects is reported. The fundamental frequency (F0) at similar places, Plosive duration at certain phoneme, Amplitude ratio (A1:A2) etc. were compared in normal and disguised speech. It was found that the formant frequency of normal and disguised speech remains almost similar only if it is compared at the position of same vowel quality and quantity. If the vowel is more closed or more open in the disguised utterance the formant frequency will be changed in comparison to normal utterance. The ratio of the amplitude (A1: A2) is found to be speaker dependent. It remains unchanged in the disguised utterance. However, this value may shift in disguised utterance if cross sectioning is not done at the same location.

  15. Evidence-based practice in speech-language pathology curricula: a scoping study.

    PubMed

    Togher, Leanne; Yiannoukas, Corina; Lincoln, Michelle; Power, Emma; Munro, Natalie; Mccabe, Patricia; Ghosh, Pratiti; Worrall, Linda; Ward, Elizabeth; Ferguson, Alison; Harrison, Elisabeth; Douglas, Jacinta

    2011-12-01

    This scoping study investigated how evidence-based practice (EBP) principles are taught in Australian speech-language pathology (SLP) teaching and learning contexts. It explored how Australian SLP university programs: (1) facilitate student learning about the principles of EBP in academic and clinical settings, and (2) self-evaluate their curricula in relation to EBP. The research involved two surveys. Survey 1 respondents were 131 academic staff, program coordinators, and on-campus and off-campus clinical educators. This survey gathered information about EBP teaching and learning in SLP programs as well as future EBP curriculum plans. Survey 2 investigated how clinical educators incorporated EBP into the way they taught clinical decision-making to students. Surveys responses from 85 clinical educators were analysed using descriptive and non-parametric statistics and thematic grouping of open-ended qualitative responses. Both surveys revealed strengths and gaps in integrating EBP into Australian SLP curricula. Perceived strengths were that respondents were positive about EBP, most had EBP training and access to EBP resources. The perceived gaps included the academic staff's perceptions of students' understanding and application of EBP, respondents' understanding of research methodologies, communication and collaboration between academic staff and clinical educators, and a lack of explicit discussion by clinical educators and students of EBP in relation to clients.

  16. A voice-input voice-output communication aid for people with severe speech impairment.

    PubMed

    Hawley, Mark S; Cunningham, Stuart P; Green, Phil D; Enderby, Pam; Palmer, Rebecca; Sehgal, Siddharth; O'Neill, Peter

    2013-01-01

    A new form of augmentative and alternative communication (AAC) device for people with severe speech impairment-the voice-input voice-output communication aid (VIVOCA)-is described. The VIVOCA recognizes the disordered speech of the user and builds messages, which are converted into synthetic speech. System development was carried out employing user-centered design and development methods, which identified and refined key requirements for the device. A novel methodology for building small vocabulary, speaker-dependent automatic speech recognizers with reduced amounts of training data, was applied. Experiments showed that this method is successful in generating good recognition performance (mean accuracy 96%) on highly disordered speech, even when recognition perplexity is increased. The selected message-building technique traded off various factors including speed of message construction and range of available message outputs. The VIVOCA was evaluated in a field trial by individuals with moderate to severe dysarthria and confirmed that they can make use of the device to produce intelligible speech output from disordered speech input. The trial highlighted some issues which limit the performance and usability of the device when applied in real usage situations, with mean recognition accuracy of 67% in these circumstances. These limitations will be addressed in future work.

  17. Speech outcomes in Parkinson's disease after subthalamic nucleus deep brain stimulation: A systematic review.

    PubMed

    Aldridge, Danielle; Theodoros, Deborah; Angwin, Anthony; Vogel, Adam P

    2016-12-01

    Deep brain stimulation (DBS) of the subthalamic nucleus (STN) is effective in reducing motor symptoms for many individuals with Parkinson's disease (PD). However, STN DBS does not appear to influence speech in the same way, and may result in a variety of negative outcomes for people with PD (PWP). A high degree of inter-individual variability amongst PWP regarding speech outcomes following STN DBS is evident in many studies. Furthermore, speech studies in PWP following STN DBS have employed a wide variety of designs and methodologies, which complicate comparison and interpretation of outcome data amongst studies within this growing body of research. An analysis of published evidence regarding speech outcomes in PWP following STN DBS, according to design and quality, is missing. This systematic review aimed to analyse and coalesce all of the current evidence reported within observational and experimental studies investigating the effects of STN DBS on speech. It will strengthen understanding of the relationship between STN DBS and speech, and inform future research by highlighting methodological limitations of current evidence. Copyright © 2016 Elsevier Ltd. All rights reserved.

  18. Speech for the National Black Child Development Institute Conference (St. Louis, Missouri, October 25, 1991).

    ERIC Educational Resources Information Center

    Clay, William L.

    This speech by a black congressman from Missouri concerns the context in which general questions about black children must be answered. These questions include: (1) How can service providers, institutions, and government entities implement a vision for the future of black children? and (2) How can organizations that serve black children ensure…

  19. Remote Capture of Human Voice Acoustical Data by Telephone: A Methods Study

    ERIC Educational Resources Information Center

    Cannizzaro, Michael S.; Reilly, Nicole; Mundt, James C.; Snyder, Peter J.

    2005-01-01

    In this pilot study we sought to determine the reliability and validity of collecting speech and voice acoustical data via telephone transmission for possible future use in large clinical trials. Simultaneous recordings of each participant's speech and voice were made at the point of participation, the local recording (LR), and over a telephone…

  20. Stuttering Frequency, Speech Rate, Speech Naturalness, and Speech Effort During the Production of Voluntary Stuttering.

    PubMed

    Davidow, Jason H; Grossman, Heather L; Edge, Robin L

    2018-05-01

    Voluntary stuttering techniques involve persons who stutter purposefully interjecting disfluencies into their speech. Little research has been conducted on the impact of these techniques on the speech pattern of persons who stutter. The present study examined whether changes in the frequency of voluntary stuttering accompanied changes in stuttering frequency, articulation rate, speech naturalness, and speech effort. In total, 12 persons who stutter aged 16-34 years participated. Participants read four 300-syllable passages during a control condition, and three voluntary stuttering conditions that involved attempting to produce purposeful, tension-free repetitions of initial sounds or syllables of a word for two or more repetitions (i.e., bouncing). The three voluntary stuttering conditions included bouncing on 5%, 10%, and 15% of syllables read. Friedman tests and follow-up Wilcoxon signed ranks tests were conducted for the statistical analyses. Stuttering frequency, articulation rate, and speech naturalness were significantly different between the voluntary stuttering conditions. Speech effort did not differ between the voluntary stuttering conditions. Stuttering frequency was significantly lower during the three voluntary stuttering conditions compared to the control condition, and speech effort was significantly lower during two of the three voluntary stuttering conditions compared to the control condition. Due to changes in articulation rate across the voluntary stuttering conditions, it is difficult to conclude, as has been suggested previously, that voluntary stuttering is the reason for stuttering reductions found when using voluntary stuttering techniques. Additionally, future investigations should examine different types of voluntary stuttering over an extended period of time to determine their impact on stuttering frequency, speech rate, speech naturalness, and speech effort.

  1. Temporal plasticity in auditory cortex improves neural discrimination of speech sounds

    PubMed Central

    Engineer, Crystal T.; Shetake, Jai A.; Engineer, Navzer D.; Vrana, Will A.; Wolf, Jordan T.; Kilgard, Michael P.

    2017-01-01

    Background Many individuals with language learning impairments exhibit temporal processing deficits and degraded neural responses to speech sounds. Auditory training can improve both the neural and behavioral deficits, though significant deficits remain. Recent evidence suggests that vagus nerve stimulation (VNS) paired with rehabilitative therapies enhances both cortical plasticity and recovery of normal function. Objective/Hypothesis We predicted that pairing VNS with rapid tone trains would enhance the primary auditory cortex (A1) response to unpaired novel speech sounds. Methods VNS was paired with tone trains 300 times per day for 20 days in adult rats. Responses to isolated speech sounds, compressed speech sounds, word sequences, and compressed word sequences were recorded in A1 following the completion of VNS-tone train pairing. Results Pairing VNS with rapid tone trains resulted in stronger, faster, and more discriminable A1 responses to speech sounds presented at conversational rates. Conclusion This study extends previous findings by documenting that VNS paired with rapid tone trains altered the neural response to novel unpaired speech sounds. Future studies are necessary to determine whether pairing VNS with appropriate auditory stimuli could potentially be used to improve both neural responses to speech sounds and speech perception in individuals with receptive language disorders. PMID:28131520

  2. Musing on the Memes of Open and Distance Education

    ERIC Educational Resources Information Center

    Latchem, Colin

    2014-01-01

    Just as genes propagate themselves in the gene pool by leaping from body to body, so memes (ideas, behaviours, and actions) transmit cultural ideas or practices from one mind to another through writing, speech, or other imitable phenomena. This paper considers the memes that influence the evolution of open and distance education. If the…

  3. Current Policies and New Directions for Speech-Language Pathology Assistants.

    PubMed

    Paul-Brown, Diane; Goldberg, Lynette R

    2001-01-01

    This article provides an overview of current American Speech-Language-Hearing Association (ASHA) policies for the appropriate use and supervision of speech-language pathology assistants with an emphasis on the need to preserve the role of fully qualified speech-language pathologists in the service delivery system. Seven challenging issues surrounding the appropriate use of speech-language pathology assistants are considered. These include registering assistants and approving training programs; membership in ASHA; discrepancies between state requirements and ASHA policies; preparation for serving diverse multicultural, bilingual, and international populations; supervision considerations; funding and reimbursement for assistants; and perspectives on career-ladder/bachelor-level personnel. The formation of a National Leadership Council is proposed to develop a coordinated strategic plan for addressing these controversial and potentially divisive issues related to speech-language pathology assistants. This council would implement strategies for future development in the areas of professional education pertaining to assistant-level supervision, instruction of assistants, communication networks, policy development, research, and the dissemination/promotion of information regarding assistants.

  4. The confidence and knowledge of health practitioners when interacting with people with aphasia in a hospital setting.

    PubMed

    Cameron, Ashley; McPhail, Steven; Hudson, Kyla; Fleming, Jennifer; Lethlean, Jennifer; Tan, Ngang Ju; Finch, Emma

    2018-06-01

    The aim of the study was to describe and compare the confidence and knowledge of health professionals (HPs) with and without specialized speech-language training for communicating with people with aphasia (PWA) in a metropolitan hospital setting. Ninety HPs from multidisciplinary teams completed a customized survey to identify their demographic information, knowledge of aphasia, current use of supported conversation strategies and overall communication confidence when interacting with PWA using a 100 mm visual analogue scale (VAS) to rate open-ended questions. Conventional descriptive statistics were used to examine the demographic information. Descriptive statistics and the Mann-Whitney U test were used to analyse VAS confidence rating data. The responses to the open-ended survey questions were grouped into four previously identified key categories. The HPs consisted of 22 (24.4%) participants who were speech-language pathologists and 68 (75.6%) participants from other disciplines (non-speech-language pathology HPs, non-SLP HPs). The non-SLP HPs reported significantly lower confidence levels (U = 159.0, p < 0.001, two-tailed) and identified fewer strategies for communicating effectively with PWA than the trained speech-language pathologists. The non-SLP HPs identified a median of two strategies identified [interquartile range (IQR) 1-3] in contrast to the speech-language pathologists who identified a median of eight strategies (IQR 7-12). These findings suggest that HPs, particularly those without specialized communication education, are likely to benefit from formal training to enhance their confidence, skills and ability to successfully communicate with PWA in their work environment. This may in turn increase the involvement of PWA in their health care decisions. Implications for Rehabilitation Interventions to remediate health professional's (particularly non-speech-language pathology health professionals) lower levels of confidence and ability to communicate with PWA may ultimately help ensure equal access for PWA. Promote informed collaborative decision-making, and foster patient-centred care within the health care setting.

  5. Prosody and Semantics Are Separate but Not Separable Channels in the Perception of Emotional Speech: Test for Rating of Emotions in Speech.

    PubMed

    Ben-David, Boaz M; Multani, Namita; Shakuf, Vered; Rudzicz, Frank; van Lieshout, Pascal H H M

    2016-02-01

    Our aim is to explore the complex interplay of prosody (tone of speech) and semantics (verbal content) in the perception of discrete emotions in speech. We implement a novel tool, the Test for Rating of Emotions in Speech. Eighty native English speakers were presented with spoken sentences made of different combinations of 5 discrete emotions (anger, fear, happiness, sadness, and neutral) presented in prosody and semantics. Listeners were asked to rate the sentence as a whole, integrating both speech channels, or to focus on one channel only (prosody or semantics). We observed supremacy of congruency, failure of selective attention, and prosodic dominance. Supremacy of congruency means that a sentence that presents the same emotion in both speech channels was rated highest; failure of selective attention means that listeners were unable to selectively attend to one channel when instructed; and prosodic dominance means that prosodic information plays a larger role than semantics in processing emotional speech. Emotional prosody and semantics are separate but not separable channels, and it is difficult to perceive one without the influence of the other. Our findings indicate that the Test for Rating of Emotions in Speech can reveal specific aspects in the processing of emotional speech and may in the future prove useful for understanding emotion-processing deficits in individuals with pathologies.

  6. A Voice Enabled Procedure Browser for the International Space Station

    NASA Technical Reports Server (NTRS)

    Rayner, Manny; Chatzichrisafis, Nikos; Hockey, Beth Ann; Farrell, Kim; Renders, Jean-Michel

    2005-01-01

    Clarissa, an experimental voice enabled procedure browser that has recently been deployed on the International Space Station (ISS), is to the best of our knowledge the first spoken dialog system in space. This paper gives background on the system and the ISS procedures, then discusses the research developed to address three key problems: grammar-based speech recognition using the Regulus toolkit; SVM based methods for open microphone speech recognition; and robust side-effect free dialogue management for handling undos, corrections and confirmations.

  7. Beef exports: looking for an opening.

    PubMed

    Chandler, T

    1997-11-08

    BSE, food safety and the veterinarian's role in certification were among matters discussed by the BVA's President, Mr Ted Chandler, at a speech in Ballymena on October 29. Speaking at the BVA's Northern Ireland dinner, at which representatives of Government, local veterinary associations, the farming community and other organisations were present, Mr Chandler emphasised the need for the Government to recognise the veterinary profession's vital contribution to food hygiene and its pivotal role in ensuring that standards are maintained. The main text of his speech is given here.

  8. A Generative Model of Speech Production in Broca’s and Wernicke’s Areas

    PubMed Central

    Price, Cathy J.; Crinion, Jenny T.; MacSweeney, Mairéad

    2011-01-01

    Speech production involves the generation of an auditory signal from the articulators and vocal tract. When the intended auditory signal does not match the produced sounds, subsequent articulatory commands can be adjusted to reduce the difference between the intended and produced sounds. This requires an internal model of the intended speech output that can be compared to the produced speech. The aim of this functional imaging study was to identify brain activation related to the internal model of speech production after activation related to vocalization, auditory feedback, and movement in the articulators had been controlled. There were four conditions: silent articulation of speech, non-speech mouth movements, finger tapping, and visual fixation. In the speech conditions, participants produced the mouth movements associated with the words “one” and “three.” We eliminated auditory feedback from the spoken output by instructing participants to articulate these words without producing any sound. The non-speech mouth movement conditions involved lip pursing and tongue protrusions to control for movement in the articulators. The main difference between our speech and non-speech mouth movement conditions is that prior experience producing speech sounds leads to the automatic and covert generation of auditory and phonological associations that may play a role in predicting auditory feedback. We found that, relative to non-speech mouth movements, silent speech activated Broca’s area in the left dorsal pars opercularis and Wernicke’s area in the left posterior superior temporal sulcus. We discuss these results in the context of a generative model of speech production and propose that Broca’s and Wernicke’s areas may be involved in predicting the speech output that follows articulation. These predictions could provide a mechanism by which rapid movement of the articulators is precisely matched to the intended speech outputs during future articulations. PMID:21954392

  9. Cochlear implant – state of the art

    PubMed Central

    Lenarz, Thomas

    2018-01-01

    Cochlear implants are the treatment of choice for auditory rehabilitation of patients with sensory deafness. They restore the missing function of inner hair cells by transforming the acoustic signal into electrical stimuli for activation of auditory nerve fibers. Due to the very fast technology development, cochlear implants provide open-set speech understanding in the majority of patients including the use of the telephone. Children can achieve a near to normal speech and language development provided their deafness is detected early after onset and implantation is performed quickly thereafter. The diagnostic procedure as well as the surgical technique have been standardized and can be adapted to the individual anatomical and physiological needs both in children and adults. Special cases such as cochlear obliteration might require special measures and re-implantation, which can be done in most cases in a straight forward way. Technology upgrades count for better performance. Future developments will focus on better electrode-nerve interfaces by improving electrode technology. An increased number of electrical contacts as well as the biological treatment with regeneration of the dendrites growing onto the electrode will increase the number of electrical channels. This will give room for improved speech coding strategies in order to create the bionic ear, i.e. to restore the process of natural hearing by means of technology. The robot-assisted surgery will allow for high precision surgery and reliable hearing preservation. Biological therapies will support the bionic ear. Methods are bio-hybrid electrodes, which are coded by stem cells transplanted into the inner ear to enhance auto-production of neurotrophins. Local drug delivery will focus on suppression of trauma reaction and local regeneration. Gene therapy by nanoparticles will hopefully lead to the preservation of residual hearing in patients being affected by genetic hearing loss. Overall the cochlear implant is a very powerful tool to rehabilitate patients with sensory deafness. More than 1 million of candidates in Germany today could benefit from this high technology auditory implant. Only 50,000 are implanted so far. In the future, the procedure can be done under local anesthesia, will be minimally invasive and straight forward. Hearing preservation will be routine. PMID:29503669

  10. The Impact of Personal and/or Close Relationship Experience on Memorable Messages about Breast Cancer and the Perceived Speech Acts of the Sender

    PubMed Central

    Smith, Sandi W.; Atkin, Charles; Skubisz, Christine M.; Munday, Samantha; Stohl, Cynthia

    2009-01-01

    Background Memorable messages and their speech acts (purposes of the messages) can promote protection against breast cancer and guide health behaviors. Methods Participants reported their personal, friends’, and relatives’ experiences with breast cancer and a memorable message about breast cancer if one came to mind. Those with a memorable message reported its perceived speech acts. Results Individuals who had personal and friend or relative experience with breast cancer were significantly more likely to recall memorable messages than other respondents. The most frequently perceived speech acts were providing facts, providing advice, and giving hope. Conclusion This information should be used to form messages in future breast cancer protection campaigns. PMID:19431030

  11. Randomized controlled trial of video self-modeling following speech restructuring treatment for stuttering.

    PubMed

    Cream, Angela; O'Brian, Sue; Jones, Mark; Block, Susan; Harrison, Elisabeth; Lincoln, Michelle; Hewat, Sally; Packman, Ann; Menzies, Ross; Onslow, Mark

    2010-08-01

    In this study, the authors investigated the efficacy of video self-modeling (VSM) following speech restructuring treatment to improve the maintenance of treatment effects. The design was an open-plan, parallel-group, randomized controlled trial. Participants were 89 adults and adolescents who undertook intensive speech restructuring treatment. Post treatment, participants were randomly assigned to 2 trial arms: standard maintenance and standard maintenance plus VSM. Participants in the latter arm viewed stutter-free videos of themselves each day for 1 month. The addition of VSM did not improve speech outcomes, as measured by percent syllables stuttered, at either 1 or 6 months postrandomization. However, at the latter assessment, self-rating of worst stuttering severity by the VSM group was 10% better than that of the control group, and satisfaction with speech fluency was 20% better. Quality of life was also better for the VSM group, which was mildly to moderately impaired compared with moderate impairment in the control group. VSM intervention after treatment was associated with improvements in self-reported outcomes. The clinical implications of this finding are discussed.

  12. The effect of sensorineural hearing loss and tinnitus on speech recognition over air and bone conduction military communications headsets.

    PubMed

    Manning, Candice; Mermagen, Timothy; Scharine, Angelique

    2017-06-01

    Military personnel are at risk for hearing loss due to noise exposure during deployment (USACHPPM, 2008). Despite mandated use of hearing protection, hearing loss and tinnitus are prevalent due to reluctance to use hearing protection. Bone conduction headsets can offer good speech intelligibility for normal hearing (NH) listeners while allowing the ears to remain open in quiet environments and the use of hearing protection when needed. Those who suffer from tinnitus, the experience of perceiving a sound not produced by an external source, often show degraded speech recognition; however, it is unclear whether this is a result of decreased hearing sensitivity or increased distractibility (Moon et al., 2015). It has been suggested that the vibratory stimulation of a bone conduction headset might ameliorate the effects of tinnitus on speech perception; however, there is currently no research to support or refute this claim (Hoare et al., 2014). Speech recognition of words presented over air conduction and bone conduction headsets was measured for three groups of listeners: NH, sensorineural hearing impaired, and/or tinnitus sufferers. Three levels of speech-to-noise (SNR = 0, -6, -12 dB) were created by embedding speech items in pink noise. Better speech recognition performance was observed with the bone conduction headset regardless of hearing profile, and speech intelligibility was a function of SNR. Discussion will include study limitations and the implications of these findings for those serving in the military. Published by Elsevier B.V.

  13. Internet Video Telephony Allows Speech Reading by Deaf Individuals and Improves Speech Perception by Cochlear Implant Users

    PubMed Central

    Mantokoudis, Georgios; Dähler, Claudia; Dubach, Patrick; Kompis, Martin; Caversaccio, Marco D.; Senn, Pascal

    2013-01-01

    Objective To analyze speech reading through Internet video calls by profoundly hearing-impaired individuals and cochlear implant (CI) users. Methods Speech reading skills of 14 deaf adults and 21 CI users were assessed using the Hochmair Schulz Moser (HSM) sentence test. We presented video simulations using different video resolutions (1280×720, 640×480, 320×240, 160×120 px), frame rates (30, 20, 10, 7, 5 frames per second (fps)), speech velocities (three different speakers), webcameras (Logitech Pro9000, C600 and C500) and image/sound delays (0–500 ms). All video simulations were presented with and without sound and in two screen sizes. Additionally, scores for live Skype™ video connection and live face-to-face communication were assessed. Results Higher frame rate (>7 fps), higher camera resolution (>640×480 px) and shorter picture/sound delay (<100 ms) were associated with increased speech perception scores. Scores were strongly dependent on the speaker but were not influenced by physical properties of the camera optics or the full screen mode. There is a significant median gain of +8.5%pts (p = 0.009) in speech perception for all 21 CI-users if visual cues are additionally shown. CI users with poor open set speech perception scores (n = 11) showed the greatest benefit under combined audio-visual presentation (median speech perception +11.8%pts, p = 0.032). Conclusion Webcameras have the potential to improve telecommunication of hearing-impaired individuals. PMID:23359119

  14. Song and speech: examining the link between singing talent and speech imitation ability.

    PubMed

    Christiner, Markus; Reiterer, Susanne M

    2013-01-01

    In previous research on speech imitation, musicality, and an ability to sing were isolated as the strongest indicators of good pronunciation skills in foreign languages. We, therefore, wanted to take a closer look at the nature of the ability to sing, which shares a common ground with the ability to imitate speech. This study focuses on whether good singing performance predicts good speech imitation. Forty-one singers of different levels of proficiency were selected for the study and their ability to sing, to imitate speech, their musical talent and working memory were tested. Results indicated that singing performance is a better indicator of the ability to imitate speech than the playing of a musical instrument. A multiple regression revealed that 64% of the speech imitation score variance could be explained by working memory together with educational background and singing performance. A second multiple regression showed that 66% of the speech imitation variance of completely unintelligible and unfamiliar language stimuli (Hindi) could be explained by working memory together with a singer's sense of rhythm and quality of voice. This supports the idea that both vocal behaviors have a common grounding in terms of vocal and motor flexibility, ontogenetic and phylogenetic development, neural orchestration and auditory memory with singing fitting better into the category of "speech" on the productive level and "music" on the acoustic level. As a result, good singers benefit from vocal and motor flexibility, productively and cognitively, in three ways. (1) Motor flexibility and the ability to sing improve language and musical function. (2) Good singers retain a certain plasticity and are open to new and unusual sound combinations during adulthood both perceptually and productively. (3) The ability to sing improves the memory span of the auditory working memory.

  15. A warning to the Brazilian Speech-Language Pathology and Audiology community about the importance of scientific and clinical activities in primary progressive aphasia.

    PubMed

    Beber, Bárbara Costa; Brandão, Lenisa; Chaves, Márcia Lorena Fagundes

    2015-01-01

    This article aims to warn the Brazilian Speech-Language Pathology and Audiology scientific community about the importance and necessity of scientific and clinical activities regarding Primary Progressive Aphasia. This warning is based on a systematic literature review of the scientific production on Primary Progressive Aphasia, from which nine Brazilian articles were selected. It was observed that there is an obvious lack of studies on the subject, as all the retrieved articles were published in medical journals and much of it consisted of small samples; only two articles described the effectiveness of speech-language therapy in patients with Primary Progressive Aphasia. A perspective for the future in the area and characteristics of Speech-Language Therapy for Primary Progressive Aphasia are discussed. As a conclusion, it is evident the need for greater action by Speech-Language Pathology and Audiology on Primary Progressive Aphasia.

  16. The effect of botulinum toxin A (Botox) injections used to treat limb spasticity on speech patterns in children with dysarthria and cerebral palsy: A report of two cases.

    PubMed

    Workinger, Marilyn Seif; Kent, Raymond D; Meilahn, Jill R

    2017-05-19

    Botulinum toxin A (Btx-A) injections are used to treat limb spasticity in children with cerebral palsy (CP) resulting in improved gross and fine motor control. This treatment has also been reported to have additional functional effects, but the effect of treatment on speech has not been reported. This report presents results of longitudinal speech evaluation of two children with CP given injections of Btx-A for treatment of limb spasticity. Speech evaluations were accomplished at baseline (date of injections) and 4- and 10-weeks post-injections. Improvements in production of consonants, loudness control, and syllables produced per breath were found. Parental survey also suggested improvements in subjects' speech production and willingness to speak outside the testing situation. Future larger studies are warranted to assess the nature of the changes observed related to Btx-A.

  17. Replacing a Missing Tooth

    MedlinePlus

    ... usually shifted, so this must be corrected. A space is often opened up and maintained for later ... with an artificial tooth serves to maintain the space and improve speech and appearance until a definitive ...

  18. Self-Administered Computer Therapy for Apraxia of Speech: Two-Period Randomized Control Trial With Crossover.

    PubMed

    Varley, Rosemary; Cowell, Patricia E; Dyson, Lucy; Inglis, Lesley; Roper, Abigail; Whiteside, Sandra P

    2016-03-01

    There is currently little evidence on effective interventions for poststroke apraxia of speech. We report outcomes of a trial of self-administered computer therapy for apraxia of speech. Effects of speech intervention on naming and repetition of treated and untreated words were compared with those of a visuospatial sham program. The study used a parallel-group, 2-period, crossover design, with participants receiving 2 interventions. Fifty participants with chronic and stable apraxia of speech were randomly allocated to 1 of 2 order conditions: speech-first condition versus sham-first condition. Period 1 design was equivalent to a randomized controlled trial. We report results for this period and profile the effect of the period 2 crossover. Period 1 results revealed significant improvement in naming and repetition only in the speech-first group. The sham-first group displayed improvement in speech production after speech intervention in period 2. Significant improvement of treated words was found in both naming and repetition, with little generalization to structurally similar and dissimilar untreated words. Speech gains were largely maintained after withdrawal of intervention. There was a significant relationship between treatment dose and response. However, average self-administered dose was modest for both groups. Future software design would benefit from incorporation of social and gaming components to boost motivation. Single-word production can be improved in chronic apraxia of speech with behavioral intervention. Self-administered computerized therapy is a promising method for delivering high-intensity speech/language rehabilitation. URL: http://orcid.org/0000-0002-1278-0601. Unique identifier: ISRCTN88245643. © 2016 American Heart Association, Inc.

  19. The benefits of remote microphone technology for adults with cochlear implants.

    PubMed

    Fitzpatrick, Elizabeth M; Séguin, Christiane; Schramm, David R; Armstrong, Shelly; Chénier, Josée

    2009-10-01

    Cochlear implantation has become a standard practice for adults with severe to profound hearing loss who demonstrate limited benefit from hearing aids. Despite the substantial auditory benefits provided by cochlear implants, many adults experience difficulty understanding speech in noisy environments and in other challenging listening conditions such as television. Remote microphone technology may provide some benefit in these situations; however, little is known about whether these systems are effective in improving speech understanding in difficult acoustic environments for this population. This study was undertaken with adult cochlear implant recipients to assess the potential benefits of remote microphone technology. The objectives were to examine the measurable and perceived benefit of remote microphone devices during television viewing and to assess the benefits of a frequency-modulated system for speech understanding in noise. Fifteen adult unilateral cochlear implant users were fit with remote microphone devices in a clinical environment. The study used a combination of direct measurements and patient perceptions to assess speech understanding with and without remote microphone technology. The direct measures involved a within-subject repeated-measures design. Direct measures of patients' speech understanding during television viewing were collected using their cochlear implant alone and with their implant device coupled to an assistive listening device. Questionnaires were administered to document patients' perceptions of benefits during the television-listening tasks. Speech recognition tests of open-set sentences in noise with and without remote microphone technology were also administered. Participants showed improved speech understanding for television listening when using remote microphone devices coupled to their cochlear implant compared with a cochlear implant alone. This benefit was documented both when listening to news and talk show recordings. Questionnaire results also showed statistically significant differences between listening with a cochlear implant alone and listening with a remote microphone device. Participants judged that remote microphone technology provided them with better comprehension, more confidence, and greater ease of listening. Use of a frequency-modulated system coupled to a cochlear implant also showed significant improvement over a cochlear implant alone for open-set sentence recognition in +10 and +5 dB signal to noise ratios. Benefits were measured during remote microphone use in focused-listening situations in a clinical setting, for both television viewing and speech understanding in noise in the audiometric sound suite. The results suggest that adult cochlear implant users should be counseled regarding the potential for enhanced speech understanding in difficult listening environments through the use of remote microphone technology.

  20. Speech-driven environmental control systems--a qualitative analysis of users' perceptions.

    PubMed

    Judge, Simon; Robertson, Zoë; Hawley, Mark; Enderby, Pam

    2009-05-01

    To explore users' experiences and perceptions of speech-driven environmental control systems (SPECS) as part of a larger project aiming to develop a new SPECS. The motivation for this part of the project was to add to the evidence base for the use of SPECS and to determine the key design specifications for a new speech-driven system from a user's perspective. Semi-structured interviews were conducted with 12 users of SPECS from around the United Kingdom. These interviews were transcribed and analysed using a qualitative method based on framework analysis. Reliability is the main influence on the use of SPECS. All the participants gave examples of occasions when their speech-driven system was unreliable; in some instances, this unreliability was reported as not being a problem (e.g., for changing television channels); however, it was perceived as a problem for more safety critical functions (e.g., opening a door). Reliability was cited by participants as the reason for using a switch-operated system as back up. Benefits of speech-driven systems focused on speech operation enabling access when other methods were not possible; quicker operation and better aesthetic considerations. Overall, there was a perception of increased independence from the use of speech-driven environmental control. In general, speech was considered a useful method of operating environmental controls by the participants interviewed; however, their perceptions regarding reliability often influenced their decision to have backup or alternative systems for certain functions.

  1. Education First: Building America's Future. The Fifth Annual State of American Education Speech, Seattle, Washington.

    ERIC Educational Resources Information Center

    Riley, Richard W.

    Public education in the United States faces many challenges. Ways in which districts are meeting these challenges are discussed in this State of American Education speech given by U.S. Secretary of Education Richard W. Riley. After providing an overview of American education, with mention of reading scores, drug use, the Hispanic dropout rate,…

  2. Above-Campus Services: Shaping the Promise of Cloud Computing for Higher Education

    ERIC Educational Resources Information Center

    Wheeler, Brad; Waggener, Shelton

    2009-01-01

    The concept of today's cloud computing may date back to 1961, when John McCarthy, retired Stanford professor and Turing Award winner, delivered a speech at MIT's Centennial. In that speech, he predicted that in the future, computing would become a "public utility." Yet for colleges and universities, the recent growth of pervasive, very high speed…

  3. The Rhetoric of Balance: An Analysis of Selected Speeches by Anwar El-Sadat.

    ERIC Educational Resources Information Center

    Littlefield, Robert S.

    Anwar el-Sadat's speaking style became a key factor in his ability to maintain a balance between the goals essential to Egypt's future and the position taken by Israelis in the settlement of the Mideast conflict. Three speeches (two addressing the Egyptian National Assembly, one the Israeli Knesset) were examined to explore the rhetorical choices…

  4. Categorical speech processing in Broca's area: an fMRI study using multivariate pattern-based analysis.

    PubMed

    Lee, Yune-Sang; Turkeltaub, Peter; Granger, Richard; Raizada, Rajeev D S

    2012-03-14

    Although much effort has been directed toward understanding the neural basis of speech processing, the neural processes involved in the categorical perception of speech have been relatively less studied, and many questions remain open. In this functional magnetic resonance imaging (fMRI) study, we probed the cortical regions mediating categorical speech perception using an advanced brain-mapping technique, whole-brain multivariate pattern-based analysis (MVPA). Normal healthy human subjects (native English speakers) were scanned while they listened to 10 consonant-vowel syllables along the /ba/-/da/ continuum. Outside of the scanner, individuals' own category boundaries were measured to divide the fMRI data into /ba/ and /da/ conditions per subject. The whole-brain MVPA revealed that Broca's area and the left pre-supplementary motor area evoked distinct neural activity patterns between the two perceptual categories (/ba/ vs /da/). Broca's area was also found when the same analysis was applied to another dataset (Raizada and Poldrack, 2007), which previously yielded the supramarginal gyrus using a univariate adaptation-fMRI paradigm. The consistent MVPA findings from two independent datasets strongly indicate that Broca's area participates in categorical speech perception, with a possible role of translating speech signals into articulatory codes. The difference in results between univariate and multivariate pattern-based analyses of the same data suggest that processes in different cortical areas along the dorsal speech perception stream are distributed on different spatial scales.

  5. A French-speaking speech-language pathology program in West Africa: transfer of training between Minority and Majority World countries.

    PubMed

    Topouzkhanian, Sylvia; Mijiyawa, Moustafa

    2013-02-01

    In West Africa, as in Majority World countries, people with a communication disability are generally cut-off from the normal development process. A long-term involvement of two partners (Orthophonistes du Monde and Handicap International) allowed the implementation in 2003 of the first speech-language pathology qualifying course in West Africa, within the Ecole Nationale des Auxiliaires Medicaux (ENAM, National School for Medical Auxiliaries) in Lome, Togo. It is a 3-year basic training (after the baccalaureate) in the only academic training centre for medical assistants in Togo. This department has a regional purpose and aims at training French-speaking African students. French speech-language pathology lecturers had to adapt their courses to the local realities they discovered in Togo. It was important to introduce and develop knowledge and skills in the students' system of reference. African speech-language pathologists have to face many challenges: creating an African speech and language therapy, introducing language disorders and their possible cure by means other than traditional therapies, and adapting all the evaluation tests and tools for speech-language pathology to each country, each culture, and each language. Creating an African speech-language pathology profession (according to its own standards) with a real influence in West Africa opens great opportunities for schooling and social and occupational integration of people with communication disabilities.

  6. The personality of past, present and future speech-language pathology students.

    PubMed

    Byrne, Nicole

    2018-03-01

    As allied health professions change over time to keep up with and reflect a rapidly changing society, it is quite possible that the people attracted to the profession may also change. If this is the case, then knowing this could be critical for future workforce marketing, training and planning. The aim was to investigate whether the personality of students entering a speech-language pathology (SLP) program had changed over time and whether there were generational differences in personality. The study used the Big Five personality inventory to consider whether there were differences in the personality in speech-language pathology (SLP) students enrolled in the same regional university in Australia in 2005 and 2016. The results showed there were significant differences between the two groups on the Agreeableness and Extroversion scales. The students who were more Conscientious were also more Confident in their ability to perform as an SLP. Generational differences across the two cohorts were also considered. SLP is a dynamic profession that is reflected through an evolving scope of practice, increasing utilization of technology and specialization. As careers evolve it is logical that the people attracted to those careers may also shift; as demonstrated here via changes in the personality of SLP students. Understanding the personality of current SLP students and future Generation Z students may assist universities to identify specific skills and experiences students need to be successful in the workforce. © 2017 Royal College of Speech and Language Therapists.

  7. Visually guided auditory attention in a dynamic "cocktail-party" speech perception task: ERP evidence for age-related differences.

    PubMed

    Getzmann, Stephan; Wascher, Edmund

    2017-02-01

    Speech understanding in the presence of concurring sound is a major challenge especially for older persons. In particular, conversational turn-takings usually result in switch costs, as indicated by declined speech perception after changes in the relevant target talker. Here, we investigated whether visual cues indicating the future position of a target talker may reduce the costs of switching in younger and older adults. We employed a speech perception task, in which sequences of short words were simultaneously presented by three talkers, and analysed behavioural measures and event-related potentials (ERPs). Informative cues resulted in increased performance after a spatial change in target talker compared to uninformative cues, not indicating the future target position. Especially the older participants benefited from knowing the future target position in advance, indicated by reduced response times after informative cues. The ERP analysis revealed an overall reduced N2, and a reduced P3b to changes in the target talker location in older participants, suggesting reduced inhibitory control and context updating. On the other hand, a pronounced frontal late positive complex (f-LPC) to the informative cues indicated increased allocation of attentional resources to changes in target talker in the older group, in line with the decline-compensation hypothesis. Thus, knowing where to listen has the potential to compensate for age-related decline in attentional switching in a highly variable cocktail-party environment. Copyright © 2016 Elsevier B.V. All rights reserved.

  8. Executive functions in mono- and bilingual children with language impairment - issues for speech-language pathology.

    PubMed

    Sandgren, Olof; Holmström, Ketty

    2015-01-01

    The clinical assessment of language impairment (LI) in bilingual children imposes challenges for speech-language pathology services. Assessment tools standardized for monolingual populations increase the risk of misinterpreting bilingualism as LI. This Perspective article summarizes recent studies on the assessment of bilingual LI and presents new results on including non-linguistic measures of executive functions in the diagnostic assessment. Executive functions shows clinical utility as less subjected to language use and exposure than linguistic measures. A possible bilingual advantage, and consequences for speech-language pathology practices and future research are discussed.

  9. Noise reduction algorithm with the soft thresholding based on the Shannon entropy and bone-conduction speech cross- correlation bands.

    PubMed

    Na, Sung Dae; Wei, Qun; Seong, Ki Woong; Cho, Jin Ho; Kim, Myoung Nam

    2018-01-01

    The conventional methods of speech enhancement, noise reduction, and voice activity detection are based on the suppression of noise or non-speech components of the target air-conduction signals. However, air-conduced speech is hard to differentiate from babble or white noise signals. To overcome this problem, the proposed algorithm uses the bone-conduction speech signals and soft thresholding based on the Shannon entropy principle and cross-correlation of air- and bone-conduction signals. A new algorithm for speech detection and noise reduction is proposed, which makes use of the Shannon entropy principle and cross-correlation with the bone-conduction speech signals to threshold the wavelet packet coefficients of the noisy speech. The proposed method can be get efficient result by objective quality measure that are PESQ, RMSE, Correlation, SNR. Each threshold is generated by the entropy and cross-correlation approaches in the decomposed bands using the wavelet packet decomposition. As a result, the noise is reduced by the proposed method using the MATLAB simulation. To verify the method feasibility, we compared the air- and bone-conduction speech signals and their spectra by the proposed method. As a result, high performance of the proposed method is confirmed, which makes it quite instrumental to future applications in communication devices, noisy environment, construction, and military operations.

  10. Effect of concurrent walking and interlocutor distance on conversational speech intensity and rate in Parkinson's disease.

    PubMed

    McCaig, Cassandra M; Adams, Scott G; Dykstra, Allyson D; Jog, Mandar

    2016-01-01

    Previous studies have demonstrated a negative effect of concurrent walking and talking on gait in Parkinson's disease (PD) but there is limited information about the effect of concurrent walking on speech production. The present study examined the effect of sitting, standing, and three concurrent walking tasks (slow, normal, fast) on conversational speech intensity and speech rate in fifteen individuals with hypophonia related to idiopathic Parkinson's disease (PD) and fourteen age-equivalent controls. Interlocuter (talker-to-talker) distance effects and walking speed were also examined. Concurrent walking was found to produce a significant increase in speech intensity, relative to standing and sitting, in both the control and PD groups. Faster walking produced significantly greater speech intensity than slower walking. Concurrent walking had no effect on speech rate. Concurrent walking and talking produced significant reductions in walking speed in both the control and PD groups. In general, the results of the present study indicate that concurrent walking tasks and the speed of concurrent walking can have a significant positive effect on conversational speech intensity. These positive, "energizing" effects need to be given consideration in future attempts to develop a comprehensive model of speech intensity regulation and they may have important implications for the development of new evaluation and treatment procedures for individuals with hypophonia related to PD. Crown Copyright © 2015. Published by Elsevier B.V. All rights reserved.

  11. Hemodynamics of speech production: An fNIRS investigation of children who stutter.

    PubMed

    Walsh, B; Tian, F; Tourville, J A; Yücel, M A; Kuczek, T; Bostian, A J

    2017-06-22

    Stuttering affects nearly 1% of the population worldwide and often has life-altering negative consequences, including poorer mental health and emotional well-being, and reduced educational and employment achievements. Over two decades of neuroimaging research reveals clear anatomical and physiological differences in the speech neural networks of adults who stutter. However, there have been few neurophysiological investigations of speech production in children who stutter. Using functional near-infrared spectroscopy (fNIRS), we examined hemodynamic responses over neural regions integral to fluent speech production including inferior frontal gyrus, premotor cortex, and superior temporal gyrus during a picture description task. Thirty-two children (16 stuttering and 16 controls) aged 7-11 years participated in the study. We found distinctly different speech-related hemodynamic responses in the group of children who stutter compared to the control group. Whereas controls showed significant activation over left dorsal inferior frontal gyrus and left premotor cortex, children who stutter exhibited deactivation over these left hemisphere regions. This investigation of neural activation during natural, connected speech production in children who stutter demonstrates that in childhood stuttering, atypical functional organization for speech production is present and suggests promise for the use of fNIRS during natural speech production in future research with typical and atypical child populations.

  12. An Intrinsically Digital Amplification Scheme for Hearing Aids

    NASA Astrophysics Data System (ADS)

    Blamey, Peter J.; Macfarlane, David S.; Steele, Brenton R.

    2005-12-01

    Results for linear and wide-dynamic range compression were compared with a new 64-channel digital amplification strategy in three separate studies. The new strategy addresses the requirements of the hearing aid user with efficient computations on an open-platform digital signal processor (DSP). The new amplification strategy is not modeled on prior analog strategies like compression and linear amplification, but uses statistical analysis of the signal to optimize the output dynamic range in each frequency band independently. Using the open-platform DSP processor also provided the opportunity for blind trial comparisons of the different processing schemes in BTE and ITE devices of a high commercial standard. The speech perception scores and questionnaire results show that it is possible to provide improved audibility for sound in many narrow frequency bands while simultaneously improving comfort, speech intelligibility in noise, and sound quality.

  13. Vibration and Noise in Magnetic Resonance Imaging of the Vocal Tract: Differences between Whole-Body and Open-Air Devices.

    PubMed

    Přibil, Jiří; Přibilová, Anna; Frollo, Ivan

    2018-04-05

    This article compares open-air and whole-body magnetic resonance imaging (MRI) equipment working with a weak magnetic field as regards the methods of its generation, spectral properties of mechanical vibration and acoustic noise produced by gradient coils during the scanning process, and the measured noise intensity. These devices are used for non-invasive MRI reconstruction of the human vocal tract during phonation with simultaneous speech recording. In this case, the vibration and noise have negative influence on quality of speech signal. Two basic measurement experiments were performed within the paper: mapping sound pressure levels in the MRI device vicinity and picking up vibration and noise signals in the MRI scanning area. Spectral characteristics of these signals are then analyzed statistically and compared visually and numerically.

  14. Automated analysis of free speech predicts psychosis onset in high-risk youths

    PubMed Central

    Bedi, Gillinder; Carrillo, Facundo; Cecchi, Guillermo A; Slezak, Diego Fernández; Sigman, Mariano; Mota, Natália B; Ribeiro, Sidarta; Javitt, Daniel C; Copelli, Mauro; Corcoran, Cheryl M

    2015-01-01

    Background/Objectives: Psychiatry lacks the objective clinical tests routinely used in other specializations. Novel computerized methods to characterize complex behaviors such as speech could be used to identify and predict psychiatric illness in individuals. AIMS: In this proof-of-principle study, our aim was to test automated speech analyses combined with Machine Learning to predict later psychosis onset in youths at clinical high-risk (CHR) for psychosis. Methods: Thirty-four CHR youths (11 females) had baseline interviews and were assessed quarterly for up to 2.5 years; five transitioned to psychosis. Using automated analysis, transcripts of interviews were evaluated for semantic and syntactic features predicting later psychosis onset. Speech features were fed into a convex hull classification algorithm with leave-one-subject-out cross-validation to assess their predictive value for psychosis outcome. The canonical correlation between the speech features and prodromal symptom ratings was computed. Results: Derived speech features included a Latent Semantic Analysis measure of semantic coherence and two syntactic markers of speech complexity: maximum phrase length and use of determiners (e.g., which). These speech features predicted later psychosis development with 100% accuracy, outperforming classification from clinical interviews. Speech features were significantly correlated with prodromal symptoms. Conclusions: Findings support the utility of automated speech analysis to measure subtle, clinically relevant mental state changes in emergent psychosis. Recent developments in computer science, including natural language processing, could provide the foundation for future development of objective clinical tests for psychiatry. PMID:27336038

  15. Speech disorders of Parkinsonism: a review.

    PubMed Central

    Critchley, E M

    1981-01-01

    Study of the speech disorders of Parkinsonism provides a paradigm of the integration of phonation, articulation and language in the production of speech. The initial defect in the untreated patient is a failure to control respiration for the purpose of speech and there follows a forward progression of articulatory symptoms involving larynx, pharynx, tongue and finally lips. There is evidence that the integration of speech production is organised asymmetrically at thalamic level. Experimental or therapeutic lesions in the region of the inferior medial portion of ventro-lateral thalamus may influence the initiation, respiratory control, rate and prosody of speech. Higher language functions may also be involved in thalamic integration: different forms of anomia are reported with pulvinar and ventrolateral thalamic lesions and transient aphasia may follow stereotaxis. The results of treatment with levodopa indicates that neurotransmitter substances enhance the clarity, volume and persistence of phonation and the latency and smoothness of articulation. The improvement of speech performance is not necessarily in phase with locomotor changes. The dose-related dyskinetic effects of levodopa, which appear to have a physiological basis in observations previously made in post-encephalitic Parkinsonism, not only influence the prosody of speech with near-mutism, hesitancy and dysfluency but may affect work-finding ability and in instances of excitement (erethism) even involve the association of long-term memory with speech. In future, neurologists will need to examine more closely the role of neurotransmitters in speech production and formulation. PMID:7031185

  16. Lip Movement Exaggerations During Infant-Directed Speech

    PubMed Central

    Green, Jordan R.; Nip, Ignatius S. B.; Wilson, Erin M.; Mefferd, Antje S.; Yunusova, Yana

    2011-01-01

    Purpose Although a growing body of literature has indentified the positive effects of visual speech on speech and language learning, oral movements of infant-directed speech (IDS) have rarely been studied. This investigation used 3-dimensional motion capture technology to describe how mothers modify their lip movements when talking to their infants. Method Lip movements were recorded from 25 mothers as they spoke to their infants and other adults. Lip shapes were analyzed for differences across speaking conditions. The maximum fundamental frequency, duration, acoustic intensity, and first and second formant frequency of each vowel also were measured. Results Lip movements were significantly larger during IDS than during adult-directed speech, although the exaggerations were vowel specific. All of the vowels produced during IDS were characterized by an elevated vocal pitch and a slowed speaking rate when compared with vowels produced during adult-directed speech. Conclusion The pattern of lip-shape exaggerations did not provide support for the hypothesis that mothers produce exemplar visual models of vowels during IDS. Future work is required to determine whether the observed increases in vertical lip aperture engender visual and acoustic enhancements that facilitate the early learning of speech. PMID:20699342

  17. Simulated learning environments in speech-language pathology: an Australian response.

    PubMed

    MacBean, Naomi; Theodoros, Deborah; Davidson, Bronwyn; Hill, Anne E

    2013-06-01

    The rising demand for health professionals to service the Australian population is placing pressure on traditional approaches to clinical education in the allied health professions. Existing research suggests that simulated learning environments (SLEs) have the potential to increase student placement capacity while providing quality learning experiences with comparable or superior outcomes to traditional methods. This project investigated the current use of SLEs in Australian speech-language pathology curricula, and the potential future applications of SLEs to the clinical education curricula through an extensive consultative process with stakeholders (all 10 Australian universities offering speech-language pathology programs in 2010, Speech Pathology Australia, members of the speech-language pathology profession, and current student body). Current use of SLEs in speech-language pathology education was found to be limited, with additional resources required to further develop SLEs and maintain their use within the curriculum. Perceived benefits included: students' increased clinical skills prior to workforce placement, additional exposure to specialized areas of speech-language pathology practice, inter-professional learning, and richer observational experiences for novice students. Stakeholders perceived SLEs to have considerable potential for clinical learning. A nationally endorsed recommendation for SLE development and curricula integration was prepared.

  18. The Use of E-supervision to Support Speech-Language Pathology Graduate Students during Student Teaching Practica

    PubMed Central

    Carlin, Charles H.; Boarman, Katie; Carlin, Emily; Inselmann, Karissa

    2013-01-01

    In the present feasibility study, e-supervision was used to provide university liaison supervision to speech-language pathology (SLP) graduate students enrolled in student teaching practica. Utilizing a mixed methodology approach, interview and survey data were compared in order to identify similarities and differences between in-person and e-supervision, and guide future practice. Results showed e-supervised graduate students perceived that they received adequate supervision, feedback, support, and communication. Further, e-supervision provided additional benefits to supervisors, children on the caseload, and universities. Despite the benefits, disadvantages emerged. Implications for future practice and limitations of the study were identified. PMID:25945215

  19. Is automatic speech-to-text transcription ready for use in psychological experiments?

    PubMed

    Ziman, Kirsten; Heusser, Andrew C; Fitzpatrick, Paxton C; Field, Campbell E; Manning, Jeremy R

    2018-04-23

    Verbal responses are a convenient and naturalistic way for participants to provide data in psychological experiments (Salzinger, The Journal of General Psychology, 61(1),65-94:1959). However, audio recordings of verbal responses typically require additional processing, such as transcribing the recordings into text, as compared with other behavioral response modalities (e.g., typed responses, button presses, etc.). Further, the transcription process is often tedious and time-intensive, requiring human listeners to manually examine each moment of recorded speech. Here we evaluate the performance of a state-of-the-art speech recognition algorithm (Halpern et al., 2016) in transcribing audio data into text during a list-learning experiment. We compare transcripts made by human annotators to the computer-generated transcripts. Both sets of transcripts matched to a high degree and exhibited similar statistical properties, in terms of the participants' recall performance and recall dynamics that the transcripts captured. This proof-of-concept study suggests that speech-to-text engines could provide a cheap, reliable, and rapid means of automatically transcribing speech data in psychological experiments. Further, our findings open the door for verbal response experiments that scale to thousands of participants (e.g., administered online), as well as a new generation of experiments that decode speech on the fly and adapt experimental parameters based on participants' prior responses.

  20. The Auditory-Brainstem Response to Continuous, Non-repetitive Speech Is Modulated by the Speech Envelope and Reflects Speech Processing

    PubMed Central

    Reichenbach, Chagit S.; Braiman, Chananel; Schiff, Nicholas D.; Hudspeth, A. J.; Reichenbach, Tobias

    2016-01-01

    The auditory-brainstem response (ABR) to short and simple acoustical signals is an important clinical tool used to diagnose the integrity of the brainstem. The ABR is also employed to investigate the auditory brainstem in a multitude of tasks related to hearing, such as processing speech or selectively focusing on one speaker in a noisy environment. Such research measures the response of the brainstem to short speech signals such as vowels or words. Because the voltage signal of the ABR has a tiny amplitude, several hundred to a thousand repetitions of the acoustic signal are needed to obtain a reliable response. The large number of repetitions poses a challenge to assessing cognitive functions due to neural adaptation. Here we show that continuous, non-repetitive speech, lasting several minutes, may be employed to measure the ABR. Because the speech is not repeated during the experiment, the precise temporal form of the ABR cannot be determined. We show, however, that important structural features of the ABR can nevertheless be inferred. In particular, the brainstem responds at the fundamental frequency of the speech signal, and this response is modulated by the envelope of the voiced parts of speech. We accordingly introduce a novel measure that assesses the ABR as modulated by the speech envelope, at the fundamental frequency of speech and at the characteristic latency of the response. This measure has a high signal-to-noise ratio and can hence be employed effectively to measure the ABR to continuous speech. We use this novel measure to show that the ABR is weaker to intelligible speech than to unintelligible, time-reversed speech. The methods presented here can be employed for further research on speech processing in the auditory brainstem and can lead to the development of future clinical diagnosis of brainstem function. PMID:27303286

  1. Modeling Driving Performance Using In-Vehicle Speech Data From a Naturalistic Driving Study.

    PubMed

    Kuo, Jonny; Charlton, Judith L; Koppel, Sjaan; Rudin-Brown, Christina M; Cross, Suzanne

    2016-09-01

    We aimed to (a) describe the development and application of an automated approach for processing in-vehicle speech data from a naturalistic driving study (NDS), (b) examine the influence of child passenger presence on driving performance, and (c) model this relationship using in-vehicle speech data. Parent drivers frequently engage in child-related secondary behaviors, but the impact on driving performance is unknown. Applying automated speech-processing techniques to NDS audio data would facilitate the analysis of in-vehicle driver-child interactions and their influence on driving performance. Speech activity detection and speaker diarization algorithms were applied to audio data from a Melbourne-based NDS involving 42 families. Multilevel models were developed to evaluate the effect of speech activity and the presence of child passengers on driving performance. Speech activity was significantly associated with velocity and steering angle variability. Child passenger presence alone was not associated with changes in driving performance. However, speech activity in the presence of two child passengers was associated with the most variability in driving performance. The effects of in-vehicle speech on driving performance in the presence of child passengers appear to be heterogeneous, and multiple factors may need to be considered in evaluating their impact. This goal can potentially be achieved within large-scale NDS through the automated processing of observational data, including speech. Speech-processing algorithms enable new perspectives on driving performance to be gained from existing NDS data, and variables that were once labor-intensive to process can be readily utilized in future research. © 2016, Human Factors and Ergonomics Society.

  2. Increased pain intensity is associated with greater verbal communication difficulty and increased production of speech and co-speech gestures.

    PubMed

    Rowbotham, Samantha; Wardy, April J; Lloyd, Donna M; Wearden, Alison; Holler, Judith

    2014-01-01

    Effective pain communication is essential if adequate treatment and support are to be provided. Pain communication is often multimodal, with sufferers utilising speech, nonverbal behaviours (such as facial expressions), and co-speech gestures (bodily movements, primarily of the hands and arms that accompany speech and can convey semantic information) to communicate their experience. Research suggests that the production of nonverbal pain behaviours is positively associated with pain intensity, but it is not known whether this is also the case for speech and co-speech gestures. The present study explored whether increased pain intensity is associated with greater speech and gesture production during face-to-face communication about acute, experimental pain. Participants (N = 26) were exposed to experimentally elicited pressure pain to the fingernail bed at high and low intensities and took part in video-recorded semi-structured interviews. Despite rating more intense pain as more difficult to communicate (t(25)  = 2.21, p =  .037), participants produced significantly longer verbal pain descriptions and more co-speech gestures in the high intensity pain condition (Words: t(25)  = 3.57, p  = .001; Gestures: t(25)  = 3.66, p =  .001). This suggests that spoken and gestural communication about pain is enhanced when pain is more intense. Thus, in addition to conveying detailed semantic information about pain, speech and co-speech gestures may provide a cue to pain intensity, with implications for the treatment and support received by pain sufferers. Future work should consider whether these findings are applicable within the context of clinical interactions about pain.

  3. 17 Ways to Say Yes: Toward Nuanced Tone of Voice in AAC and Speech Technology

    PubMed Central

    Pullin, Graham; Hennig, Shannon

    2015-01-01

    Abstract People with complex communication needs who use speech-generating devices have very little expressive control over their tone of voice. Despite its importance in human interaction, the issue of tone of voice remains all but absent from AAC research and development however. In this paper, we describe three interdisciplinary projects, past, present and future: The critical design collection Six Speaking Chairs has provoked deeper discussion and inspired a social model of tone of voice; the speculative concept Speech Hedge illustrates challenges and opportunities in designing more expressive user interfaces; the pilot project Tonetable could enable participatory research and seed a research network around tone of voice. We speculate that more radical interactions might expand frontiers of AAC and disrupt speech technology as a whole. PMID:25965913

  4. The Seattle Conference: The Role of the State Department of Education in Teacher Education (April 27-29, 1966).

    ERIC Educational Resources Information Center

    Edelfelt, Roy A.; Allen, Wendell C.

    This book includes the papers and speeches which were part of the Seattle Conference in which representatives of 49 states met to reevaluate beliefs and procedures regarding the future role of state departments of education particularly with reference to the identification and setting of goals. There are seven speeches and papers: "The Dynamic Duo…

  5. Keynote: FarNet Ten Years On--The Past, Present, and Future for Distance Learners

    ERIC Educational Resources Information Center

    Alexander-Bennett, Carolyn

    2016-01-01

    This think piece by Carolyn Alexander-Bennett is a reflection of her keynote speech at DEANZ2016 conference, which was held from 17-20th April at the University of Waikato, New Zealand. In her speech Carolyn revisits the issues, developments, and technology trends that led to the birth of FarNet (an online cluster of schools catering for the…

  6. Setting up a cohort study in speech and language therapy: lessons from The UK Cleft Collective Speech and Language (CC-SL) study.

    PubMed

    Wren, Yvonne; Humphries, Kerry; Stock, Nicola Marie; Rumsey, Nichola; Lewis, Sarah; Davies, Amy; Bennett, Rhiannon; Sandy, Jonathan

    2018-05-01

    Efforts to increase the evidence base in speech and language therapy are often limited by methodological factors that have restricted the strength of the evidence to the lower levels of the evidence hierarchy. Where higher graded studies, such as randomized controlled trials, have been carried out, it has sometimes been difficult to obtain sufficient power to detect a potential effect of intervention owing to small sample sizes or heterogeneity in the participants. With certain clinical groups such as cleft lip and palate, systematic reviews of intervention studies have shown that there is no robust evidence to support the efficacy of any one intervention protocol over another. To describe the setting up of an observational clinical cohort study and to present this as an alternative design for answering research questions relating to prevalence, risk factors and outcomes from intervention. The Cleft Collective Speech and Language (CC-SL) study is a national cohort study of children born with cleft palate. Working in partnership with regional clinical cleft centres, a sample size of over 600 children and 600 parents is being recruited and followed up from birth to age 5 years. Variables being collected include demographic, psychological, surgical, hearing, and speech and language data. The process of setting up the study has led to the creation of a unique, large-scale data set which is available for researchers to access now and in future. As well as exploring predictive factors, the data can be used to explore the impact of interventions in relation to individual differences. Findings from these investigations can be used to provide information on sample criteria and definitions of intervention and dosage which can be used in future trials. The observational cohort study is a useful alternative design to explore questions around prevalence, risk factors and intervention for clinical groups where robust research data are not yet available. Findings from such a study can be used to guide service-delivery decisions and to determine power for future clinical trials. © 2017 Royal College of Speech and Language Therapists.

  7. Methods and Applications of the Audibility Index in Hearing Aid Selection and Fitting

    PubMed Central

    Amlani, Amyn M.; Punch, Jerry L.; Ching, Teresa Y. C.

    2002-01-01

    During the first half of the 20th century, communications engineers at Bell Telephone Laboratories developed the articulation model for predicting speech intelligibility transmitted through different telecommunication devices under varying electroacoustic conditions. The profession of audiology adopted this model and its quantitative aspects, known as the Articulation Index and Speech Intelligibility Index, and applied these indices to the prediction of unaided and aided speech intelligibility in hearing-impaired listeners. Over time, the calculation methods of these indices—referred to collectively in this paper as the Audibility Index—have been continually refined and simplified for clinical use. This article provides (1) an overview of the basic principles and the calculation methods of the Audibility Index, the Speech Transmission Index and related indices, as well as the Speech Recognition Sensitivity Model, (2) a review of the literature on using the Audibility Index to predict speech intelligibility of hearing-impaired listeners, (3) a review of the literature on the applicability of the Audibility Index to the selection and fitting of hearing aids, and (4) a discussion of future scientific needs and clinical applications of the Audibility Index. PMID:25425917

  8. Evaluation of speech errors in Putonghua speakers with cleft palate: a critical review of methodology issues.

    PubMed

    Jiang, Chenghui; Whitehill, Tara L

    2014-04-01

    Speech errors associated with cleft palate are well established for English and several other Indo-European languages. Few articles describing the speech of Putonghua (standard Mandarin Chinese) speakers with cleft palate have been published in English language journals. Although methodological guidelines have been published for the perceptual speech evaluation of individuals with cleft palate, there has been no critical review of methodological issues in studies of Putonghua speakers with cleft palate. A literature search was conducted to identify relevant studies published over the past 30 years in Chinese language journals. Only studies incorporating perceptual analysis of speech were included. Thirty-seven articles which met inclusion criteria were analyzed and coded on a number of methodological variables. Reliability was established by having all variables recoded for all studies. This critical review identified many methodological issues. These design flaws make it difficult to draw reliable conclusions about characteristic speech errors in this group of speakers. Specific recommendations are made to improve the reliability and validity of future studies, as well to facilitate cross-center comparisons.

  9. Inferring imagined speech using EEG signals: a new approach using Riemannian manifold features

    NASA Astrophysics Data System (ADS)

    Nguyen, Chuong H.; Karavas, George K.; Artemiadis, Panagiotis

    2018-02-01

    Objective. In this paper, we investigate the suitability of imagined speech for brain-computer interface (BCI) applications. Approach. A novel method based on covariance matrix descriptors, which lie in Riemannian manifold, and the relevance vector machines classifier is proposed. The method is applied on electroencephalographic (EEG) signals and tested in multiple subjects. Main results. The method is shown to outperform other approaches in the field with respect to accuracy and robustness. The algorithm is validated on various categories of speech, such as imagined pronunciation of vowels, short words and long words. The classification accuracy of our methodology is in all cases significantly above chance level, reaching a maximum of 70% for cases where we classify three words and 95% for cases of two words. Significance. The results reveal certain aspects that may affect the success of speech imagery classification from EEG signals, such as sound, meaning and word complexity. This can potentially extend the capability of utilizing speech imagery in future BCI applications. The dataset of speech imagery collected from total 15 subjects is also published.

  10. Song and speech: examining the link between singing talent and speech imitation ability

    PubMed Central

    Christiner, Markus; Reiterer, Susanne M.

    2013-01-01

    In previous research on speech imitation, musicality, and an ability to sing were isolated as the strongest indicators of good pronunciation skills in foreign languages. We, therefore, wanted to take a closer look at the nature of the ability to sing, which shares a common ground with the ability to imitate speech. This study focuses on whether good singing performance predicts good speech imitation. Forty-one singers of different levels of proficiency were selected for the study and their ability to sing, to imitate speech, their musical talent and working memory were tested. Results indicated that singing performance is a better indicator of the ability to imitate speech than the playing of a musical instrument. A multiple regression revealed that 64% of the speech imitation score variance could be explained by working memory together with educational background and singing performance. A second multiple regression showed that 66% of the speech imitation variance of completely unintelligible and unfamiliar language stimuli (Hindi) could be explained by working memory together with a singer's sense of rhythm and quality of voice. This supports the idea that both vocal behaviors have a common grounding in terms of vocal and motor flexibility, ontogenetic and phylogenetic development, neural orchestration and auditory memory with singing fitting better into the category of “speech” on the productive level and “music” on the acoustic level. As a result, good singers benefit from vocal and motor flexibility, productively and cognitively, in three ways. (1) Motor flexibility and the ability to sing improve language and musical function. (2) Good singers retain a certain plasticity and are open to new and unusual sound combinations during adulthood both perceptually and productively. (3) The ability to sing improves the memory span of the auditory working memory. PMID:24319438

  11. The Psychologist as an Interlocutor in Autism Spectrum Disorder Assessment: Insights From a Study of Spontaneous Prosody

    PubMed Central

    Bone, Daniel; Lee, Chi-Chun; Black, Matthew P.; Williams, Marian E.; Lee, Sungbok; Levitt, Pat; Narayanan, Shrikanth

    2015-01-01

    Purpose The purpose of this study was to examine relationships between prosodic speech cues and autism spectrum disorder (ASD) severity, hypothesizing a mutually interactive relationship between the speech characteristics of the psychologist and the child. The authors objectively quantified acoustic-prosodic cues of the psychologist and of the child with ASD during spontaneous interaction, establishing a methodology for future large-sample analysis. Method Speech acoustic-prosodic features were semiautomatically derived from segments of semistructured interviews (Autism Diagnostic Observation Schedule, ADOS; Lord, Rutter, DiLavore, & Risi, 1999; Lord et al., 2012) with 28 children who had previously been diagnosed with ASD. Prosody was quantified in terms of intonation, volume, rate, and voice quality. Research hypotheses were tested via correlation as well as hierarchical and predictive regression between ADOS severity and prosodic cues. Results Automatically extracted speech features demonstrated prosodic characteristics of dyadic interactions. As rated ASD severity increased, both the psychologist and the child demonstrated effects for turn-end pitch slope, and both spoke with atypical voice quality. The psychologist’s acoustic cues predicted the child’s symptom severity better than did the child’s acoustic cues. Conclusion The psychologist, acting as evaluator and interlocutor, was shown to adjust his or her behavior in predictable ways based on the child’s social-communicative impairments. The results support future study of speech prosody of both interaction partners during spontaneous conversation, while using automatic computational methods that allow for scalable analysis on much larger corpora. PMID:24686340

  12. NREL: Speeches - Nation's Energy Future at Risk

    Science.gov Websites

    Energy Future at Risk, National Lab Director Says For more information contact: George Douglas, 303 -275-4096 e:mail: George Douglas Washington, D.C., July 27, 1999 — America must invest in its energy future now, Richard Truly, director of the U.S. Department of Energy's National Renewable Energy

  13. Comparison of mapping quantitative theta encephalograms during directed and required visual-verbal activity and passive period in children with different disorders of speech-language functioning.

    PubMed

    Radicevic, Zoran; Jelicic Dobrijevic, Ljiljana; Sovilj, Mirjana; Barlov, Ivana

    2009-06-01

    Aim of the research was to examine similarities and differences between the periods of experiencing visually stimulated directed speech-language information and periods of undirected attention. The examined group comprised N = 64 children, aged 4-5, with different speech-language disorders (developmental dysphasia, hyperactive syndrome with attention disorder, children with borderline intellectual abilities, autistic complex). Theta EEG was registered in children in the period of watching and describing the picture ("task"), and in the period of undirected attention ("passive period"). The children were recorded in standard EEG conditions, at 19 points of EEG registration and in longitudinal bipolar montage. Results in the observed age-operative theta rhythm indicated significant similarities and differences in the prevalence of spatial engagement of certain regions between the two hemispheres at the input and output of processing, which opens the possibility for more detailed analysis of conscious control of speech-language processing and its disorders.

  14. Phenomenology of future-oriented mind-wandering episodes

    PubMed Central

    Stawarczyk, David; Cassol, Helena; D'Argembeau, Arnaud

    2013-01-01

    Recent research suggests that prospective and non-prospective forms of mind-wandering possess distinct properties, yet little is known about what exactly differentiates between future-oriented and non-future-oriented mind-wandering episodes. In the present study, we used multilevel exploratory factor analyses (MEFA) to examine the factorial structure of various phenomenological dimensions of mind-wandering, and we then investigated whether future-oriented mind-wandering episodes differ from other classes of mind-wandering along the identified factors. We found that the phenomenological dimensions of mind-wandering are structured in four factors: representational format (inner speech vs. visual imagery), personal relevance, realism/concreteness, and structuration. Prospective mind-wandering differed from non-prospective mind-wandering along each of these factors. Specifically, future-oriented mind-wandering episodes involved inner speech to a greater extent, were more personally relevant, more realistic/concrete, and more often part of structured sequences of thoughts. These results show that future-oriented mind-wandering possesses a unique phenomenological signature and provide new insights into how this particular form of mind-wandering may adaptively contribute to autobiographical planning. PMID:23882236

  15. Shared acoustic codes underlie emotional communication in music and speech-Evidence from deep transfer learning.

    PubMed

    Coutinho, Eduardo; Schuller, Björn

    2017-01-01

    Music and speech exhibit striking similarities in the communication of emotions in the acoustic domain, in such a way that the communication of specific emotions is achieved, at least to a certain extent, by means of shared acoustic patterns. From an Affective Sciences points of view, determining the degree of overlap between both domains is fundamental to understand the shared mechanisms underlying such phenomenon. From a Machine learning perspective, the overlap between acoustic codes for emotional expression in music and speech opens new possibilities to enlarge the amount of data available to develop music and speech emotion recognition systems. In this article, we investigate time-continuous predictions of emotion (Arousal and Valence) in music and speech, and the Transfer Learning between these domains. We establish a comparative framework including intra- (i.e., models trained and tested on the same modality, either music or speech) and cross-domain experiments (i.e., models trained in one modality and tested on the other). In the cross-domain context, we evaluated two strategies-the direct transfer between domains, and the contribution of Transfer Learning techniques (feature-representation-transfer based on Denoising Auto Encoders) for reducing the gap in the feature space distributions. Our results demonstrate an excellent cross-domain generalisation performance with and without feature representation transfer in both directions. In the case of music, cross-domain approaches outperformed intra-domain models for Valence estimation, whereas for Speech intra-domain models achieve the best performance. This is the first demonstration of shared acoustic codes for emotional expression in music and speech in the time-continuous domain.

  16. Toward Emotionally Accessible Massive Open Online Courses (MOOCs).

    PubMed

    Hillaire, Garron; Iniesto, Francisco; Rienties, Bart

    2017-01-01

    This paper outlines an approach to evaluating the emotional content of three Massive Open Online Courses (MOOCs) using the affective computing approach of prosody detection on two different text-to-speech voices in conjunction with human raters judging the emotional content of course text. The intent of this work is to establish the potential variation on the emotional delivery of MOOC material through synthetic voice.

  17. Towards fully analog hardware reservoir computing for speech recognition

    NASA Astrophysics Data System (ADS)

    Smerieri, Anteo; Duport, François; Paquot, Yvan; Haelterman, Marc; Schrauwen, Benjamin; Massar, Serge

    2012-09-01

    Reservoir computing is a very recent, neural network inspired unconventional computation technique, where a recurrent nonlinear system is used in conjunction with a linear readout to perform complex calculations, leveraging its inherent internal dynamics. In this paper we show the operation of an optoelectronic reservoir computer in which both the nonlinear recurrent part and the readout layer are implemented in hardware for a speech recognition application. The performance obtained is close to the one of to state-of-the-art digital reservoirs, while the analog architecture opens the way to ultrafast computation.

  18. Research Priorities in Spasmodic Dysphonia

    PubMed Central

    Ludlow, Christy L.; Adler, Charles H.; Berke, Gerald S.; Bielamowicz, Steven A.; Blitzer, Andrew; Bressman, Susan B.; Hallett, Mark; Jinnah, H. A.; Juergens, Uwe; Martin, Sandra B.; Perlmutter, Joel S.; Sapienza, Christine; Singleton, Andrew; Tanner, Caroline M.; Woodson, Gayle E.

    2009-01-01

    OBJECTIVE To identify research priorities for increasing understanding of the pathogenesis, diagnosis and improved treatment of spasmodic dysphonia. STUDY DESIGN AND SETTING A multidisciplinary working group was formed including both scientists and clinicians from multiple disciplines, otolaryngology, neurology, speech pathology, genetics and neuroscience, to review currently available information on spasmodic dysphonia and to identify research priorities. RESULTS Operational definitions for spasmodic dysphonia at different levels of certainty were recommended for diagnosis and recommendations made for a multi-center multidisciplinary validation study. CONCLUSIONS The highest priority is to characterize the disorder and identify risk factors that may contribute to its onset. Future research should compare and contrast spasmodic dysphonia with other forms of focal dystonia. Development of animal models is recommended to explore hypotheses related to pathogenesis. Improved understanding of the pathophysiology of SD should provide the basis for developing new treatment options and exploratory clinical trials. SIGNIFICANCE This document should foster future research to improve the care of patients with this chronic debilitating voice and speech disorder by otolaryngology, neurology, and speech pathology. PMID:18922334

  19. Systematic Studies of Modified Vocalization: Speech Production Changes During a Variation of Metronomic Speech in Persons Who Do and Do Not Stutter

    PubMed Central

    Davidow, Jason H.; Bothe, Anne K.; Ye, Jun

    2011-01-01

    The most common way to induce fluency using rhythm requires persons who stutter to speak one syllable or one word to each beat of a metronome, but stuttering can also be eliminated when the stimulus is of a particular duration (e.g., 1 s). The present study examined stuttering frequency, speech production changes, and speech naturalness during rhythmic speech that alternated 1 s of reading with 1 s of silence. A repeated-measures design was used to compare data obtained during a control reading condition and during rhythmic reading in 10 persons who stutter (PWS) and 10 normally fluent controls. Ratings for speech naturalness were also gathered from naïve listeners. Results showed that mean vowel duration increased significantly, and the percentage of short phonated intervals decreased significantly, for both groups from the control to the experimental condition. Mean phonated interval length increased significantly for the fluent controls. Mean speech naturalness ratings during the experimental condition were approximately 7 on a 1–9 scale (1 = highly natural; 9 = highly unnatural), and these ratings were significantly correlated with vowel duration and phonated intervals for PWS. The findings indicate that PWS may be altering vocal fold vibration duration to obtain fluency during this rhythmic speech style, and that vocal fold vibration duration may have an impact on speech naturalness during rhythmic speech. Future investigations should examine speech production changes and speech naturalness during variations of this rhythmic condition. Educational Objectives The reader will be able to: (1) describe changes (from a control reading condition) in speech production variables when alternating between 1 s of reading and 1 s of silence, (2) describe which rhythmic conditions have been found to sound and feel the most natural, (3) describe methodological issues for studies about alterations in speech production variables during fluency-inducing conditions, and (4) describe which fluency-inducing conditions have been shown to involve a reduction in short phonated intervals. PMID:21664528

  20. The effect of speech rate on stuttering frequency, phonated intervals, speech effort, and speech naturalness during chorus reading.

    PubMed

    Davidow, Jason H; Ingham, Roger J

    2013-01-01

    This study examined the effect of speech rate on phonated intervals (PIs), in order to test whether a reduction in the frequency of short PIs is an important part of the fluency-inducing mechanism of chorus reading. The influence of speech rate on stuttering frequency, speaker-judged speech effort, and listener-judged naturalness was also examined. An added purpose was to determine if chorus reading could be further refined so as to provide a perceptual guide for gauging the level of physical effort exerted during speech production. A repeated-measures design was used to compare data obtained during control reading conditions and during several chorus reading conditions produced at different speech rates. Participants included 8 persons who stutter (PWS) between the ages of 16 and 32 years. There were significant reductions in the frequency of short PIs from the habitual reading condition during slower chorus conditions, no change when speech rates were matched between habitual reading and chorus conditions, and an increase in the frequency of short PIs during chorus reading produced at a faster rate than the habitual condition. Speech rate did not have an effect on stuttering frequency during chorus reading. In general, speech effort ratings improved and naturalness ratings worsened as speech rate decreased. These results provide evidence that (a) a reduction in the frequency of short PIs is not necessary for fluency improvement during chorus reading, and (b) speech rate may be altered to provide PWS with a more appropriate reference for how physically effortful normally fluent speech production should be. Future investigations should examine the necessity of changes in the activation of neural regions during chorus reading, the possibility of defining individualized units on a 9-point effort scale, and if there are upper and lower speech rate boundaries for receiving ratings of "highly natural sounding" speech during chorus reading. The reader will be able to: (1) describe the effect of changes in speech rate on the frequency of short phonated intervals during chorus reading, (2) describe changes to speaker-judged speech effort as speech rate changes during chorus reading, (3) and describe the effect of changes in speech rate on listener-judged naturalness ratings during chorus reading. Copyright © 2012 Elsevier Inc. All rights reserved.

  1. Ageing without hearing loss or cognitive impairment causes a decrease in speech intelligibility only in informational maskers.

    PubMed

    Rajan, R; Cainer, K E

    2008-06-23

    In most everyday settings, speech is heard in the presence of competing sounds and understanding speech requires skills in auditory streaming and segregation, followed by identification and recognition, of the attended signals. Ageing leads to difficulties in understanding speech in noisy backgrounds. In addition to age-related changes in hearing-related factors, cognitive factors also play a role but it is unclear to what extent these are generalized or modality-specific cognitive factors. We examined how ageing in normal-hearing decade age cohorts from 20 to 69 years affected discrimination of open-set speech in background noise. We used two types of sentences of similar structural and linguistic characteristics but different masking levels (i.e. differences in signal-to-noise ratios required for detection of sentences in a standard masker) so as to vary sentence demand, and two background maskers (one causing purely energetic masking effects and the other causing energetic and informational masking) to vary load conditions. There was a decline in performance (measured as speech reception thresholds for perception of sentences in noise) in the oldest cohort for both types of sentences, but only in the presence of the more demanding informational masker. We interpret these results to indicate a modality-specific decline in cognitive processing, likely a decrease in the ability to use acoustic and phonetic cues efficiently to segregate speech from background noise, in subjects aged >60.

  2. Speech-language pathology findings in patients with mouth breathing: multidisciplinary diagnosis according to etiology.

    PubMed

    Junqueira, Patrícia; Marchesan, Irene Queiroz; de Oliveira, Luciana Regina; Ciccone, Emílio; Haddad, Leonardo; Rizzo, Maria Cândida

    2010-11-01

    The purpose of this study was to identify and compare the results of the findings from speech-language pathology evaluations for orofacial function including tongue and lip rest postures, tonus, articulation and speech, voice and language, chewing, and deglutition in children who had a history of mouth breathing. The diagnoses for mouth breathing included: allergic rhinitis, adenoidal hypertrophy, allergic rhinitis with adenoidal hypertrophy; and/or functional mouth breathing. This study was conducted with on 414 subjects of both genders, from 2 to 16-years old. A team consisting of 3 speech-language pathologists, 1 pediatrician, 1 allergist, and 1 otolaryngologist, evaluated the patients. Multidisciplinary clinical examinations were carried out (complete blood counting, X-rays, nasofibroscopy, audiometry). The two most commonly found etiologies were allergic rhinitis, followed by functional mouth breathing. Of the 414 patients in the study, 346 received a speech-language pathology evaluation. The most prevalent finding in this group of 346 subjects was the presence of orofacial myofunctional disorders. The most frequently orofacial myofunctional disorder identified in these subjects who also presented mouth breathing included: habitual open lips rest posture, low and forward tongue rest posture and lack of adequate muscle tone. There were also no statistically significant relationships identified between etiology and speech-language diagnosis. Therefore, the specific type of etiology of mouth breathing does not appear to contribute to the presence, type, or number of speech-language findings which may result from mouth breathing behavior.

  3. Visual contribution to the multistable perception of speech.

    PubMed

    Sato, Marc; Basirat, Anahita; Schwartz, Jean-Luc

    2007-11-01

    The multistable perception of speech, or verbal transformation effect, refers to perceptual changes experienced while listening to a speech form that is repeated rapidly and continuously. In order to test whether visual information from the speaker's articulatory gestures may modify the emergence and stability of verbal auditory percepts, subjects were instructed to report any perceptual changes during unimodal, audiovisual, and incongruent audiovisual presentations of distinct repeated syllables. In a first experiment, the perceptual stability of reported auditory percepts was significantly modulated by the modality of presentation. In a second experiment, when audiovisual stimuli consisting of a stable audio track dubbed with a video track that alternated between congruent and incongruent stimuli were presented, a strong correlation between the timing of perceptual transitions and the timing of video switches was found. Finally, a third experiment showed that the vocal tract opening onset event provided by the visual input could play the role of a bootstrap mechanism in the search for transformations. Altogether, these results demonstrate the capacity of visual information to control the multistable perception of speech in its phonetic content and temporal course. The verbal transformation effect thus provides a useful experimental paradigm to explore audiovisual interactions in speech perception.

  4. Synthesized speech rate and pitch effects on intelligibility of warning messages for pilots

    NASA Technical Reports Server (NTRS)

    Simpson, C. A.; Marchionda-Frost, K.

    1984-01-01

    In civilian and military operations, a future threat-warning system with a voice display could warn pilots of other traffic, obstacles in the flight path, and/or terrain during low-altitude helicopter flights. The present study was conducted to learn whether speech rate and voice pitch of phoneme-synthesized speech affects pilot accuracy and response time to typical threat-warning messages. Helicopter pilots engaged in an attention-demanding flying task and listened for voice threat warnings presented in a background of simulated helicopter cockpit noise. Performance was measured by flying-task performance, threat-warning intelligibility, and response time. Pilot ratings were elicited for the different voice pitches and speech rates. Significant effects were obtained only for response time and for pilot ratings, both as a function of speech rate. For the few cases when pilots forgot to respond to a voice message, they remembered 90 percent of the messages accurately when queried for their response 8 to 10 sec later.

  5. A comparative analysis of whispered and normally phonated speech using an LPC-10 vocoder

    NASA Astrophysics Data System (ADS)

    Wilson, J. B.; Mosko, J. D.

    1985-12-01

    The determination of the performance of an LPC-10 vocoder in the processing of adult male and female whispered and normally phonated connected speech was the focus of this study. The LPC-10 vocoder's analysis of whispered speech compared quite favorably with similar studies which used sound spectrographic processing techniques. Shifting from phonated speech to whispered speech caused a substantial increase in the phonomic formant frequencies and formant bandwidths for both male and female speakers. The data from this study showed no evidence that the LPC-10 vocoder's ability to process voices with pitch extremes and quality extremes was limited in any significant manner. A comparison of the unprocessed natural vowel waveforms and qualities with the synthesized vowel waveforms and qualities revealed almost imperceptible differences. An LPC-10 vocoder's ability to process linguistic and dialectical suprasegmental features such as intonation, rate and stress at low bit rates should be a critical issue of concern for future research.

  6. The demand for speech pathology services for children: Do we need more or just different?

    PubMed

    Reilly, Sheena; Harper, Megan; Goldfeld, Sharon

    2016-12-01

    An inability or difficulty communicating can have a profound impact on a child's future ability to participate in society as a productive adult. Over the past few years the number of interventions for children with speech and language problems has almost doubled; the majority are targeted interventions delivered by speech pathologists. In this paper we examine the distribution of speech pathology services in metropolitan Melbourne and how these are aligned with need as defined by vulnerability in language and social disadvantage. We identified three times as many private sector services compared to public services for the 0-5 year age group. Overall there was poorer availability of services in some of the most vulnerable areas. The profound and long-term impact of impoverished childhood language, coupled with the considerable limitations on public spending, provide a strong impetus to deliver more equitably distributed speech pathology services. © 2016 Paediatrics and Child Health Division (The Royal Australasian College of Physicians).

  7. Control of complex motor gestures: orofacial muscle responses to load perturbations of lip during speech.

    PubMed

    Abbs, J H; Gracco, V L

    1984-04-01

    The contribution of ascending afferents to the control of speech movement was evaluated by applying unanticipated loads to the lower lip during the generation of combined upper lip-lower lip speech gestures. To eliminate potential contamination due to anticipation or adaptation, loads were applied randomly on only 10-15% of the trials. Physical characteristics of the perturbations were within the normal range of forces and movements involved in natural lip actions for speech. Compensatory responses in multiple facial muscles and lip movements were observed the first time a load was introduced, and achievement of the multimovement speech goals was never disrupted by these perturbations. Muscle responses were seen in the lower lip muscles, implicating corrective, feedback processes. Additionally, compensatory responses to these lower lip loads were also observed in the independently controlled muscles of the upper lip, reflecting the parallel operation of open-loop, sensorimotor mechanisms. Compensatory responses from both the upper and lower lip muscles were observed with small (1 mm) as well as large (15 mm) perturbations. The latencies of these compensatory responses were not discernible by conventional ensemble averaging. Moreover, responses at latencies of lower brain stem-mediated reflexes (i.e., 10-18 ms) were not apparent with inspection of individual records. Response latencies were determined on individual loaded trials through the use of a computer algorithm that took into account the variability of electromyograms (EMG) among the control trials. These latency measures confirmed the absence of brain stem-mediated responses and yielded response latencies that ranged from 22 to 75 ms. Response latencies appeared to be influenced by the time relation between load onset and the initiation of muscle activation. Examination of muscle activity changes for individual loaded trials revealed complementary variations in the magnitude of responses among multiple muscles contributing to a movement compensation. These observations may have implications for limb movement control if multimovement speech gestures are considered analogous to a limb action requiring coordinated movements around multiple joints. In this context, these speech motor control data might be interpreted to suggest that for complex movements, both corrective feedback and open-loop predictive processes are operating, with the latter involved in the control of coordination among multiple movement subcomponents.

  8. Using leap motion to investigate the emergence of structure in speech and language.

    PubMed

    Eryilmaz, Kerem; Little, Hannah

    2017-10-01

     In evolutionary linguistics, experiments using artificial signal spaces are being used to investigate the emergenceof speech structure. These signal spaces need to be continuous, non-discretized spaces from which discrete unitsand patterns can emerge. They need to be dissimilar from-but comparable with-the vocal tract, in order tominimize interference from pre-existing linguistic knowledge, while informing us about language. This is a hardbalance to strike. This article outlines a new approach that uses the Leap Motion, an infrared controller that canconvert manual movement in 3d space into sound. The signal space using this approach is more flexible than signalspaces in previous attempts. Further, output data using this approach is simpler to arrange and analyze. Theexperimental interface was built using free, and mostly open- source libraries in Python. We provide our sourcecode for other researchers as open source.

  9. Present Vision--Future Vision.

    ERIC Educational Resources Information Center

    Fitterman, L. Jeffrey

    This paper addresses issues of current and future technology use for and by individuals with visual impairments and blindness in Florida. Present technology applications used in vision programs in Florida are individually described, including video enlarging, speech output, large inkprint, braille print, paperless braille, and tactual output…

  10. Future Watch: Our Schools in the 21st Century.

    ERIC Educational Resources Information Center

    Montgomery, Judith K.; Herer, Gilbert R.

    1994-01-01

    This article reviews major social, technological, economic, and political trends in the United States and relates this larger perspective to the practices of speech language pathologists and audiologists in the schools. Implications of these trends for alternative futures are drawn. (Author/DB)

  11. Test-retest reliability of speech-evoked auditory brainstem response in healthy children at a low sensation level.

    PubMed

    Zakaria, Mohd Normani; Jalaei, Bahram

    2017-11-01

    Auditory brainstem responses evoked by complex stimuli such as speech syllables have been studied in normal subjects and subjects with compromised auditory functions. The stability of speech-evoked auditory brainstem response (speech-ABR) when tested over time has been reported but the literature is limited. The present study was carried out to determine the test-retest reliability of speech-ABR in healthy children at a low sensation level. Seventeen healthy children (6 boys, 11 girls) aged from 5 to 9 years (mean = 6.8 ± 3.3 years) were tested in two sessions separated by a 3-month period. The stimulus used was a 40-ms syllable /da/ presented at 30 dB sensation level. As revealed by pair t-test and intra-class correlation (ICC) analyses, peak latencies, peak amplitudes and composite onset measures of speech-ABR were found to be highly replicable. Compared to other parameters, higher ICC values were noted for peak latencies of speech-ABR. The present study was the first to report the test-retest reliability of speech-ABR recorded at low stimulation levels in healthy children. Due to its good stability, it can be used as an objective indicator for assessing the effectiveness of auditory rehabilitation in hearing-impaired children in future studies. Copyright © 2017 Elsevier B.V. All rights reserved.

  12. From Birdsong to Human Speech Recognition: Bayesian Inference on a Hierarchy of Nonlinear Dynamical Systems

    PubMed Central

    Yildiz, Izzet B.; von Kriegstein, Katharina; Kiebel, Stefan J.

    2013-01-01

    Our knowledge about the computational mechanisms underlying human learning and recognition of sound sequences, especially speech, is still very limited. One difficulty in deciphering the exact means by which humans recognize speech is that there are scarce experimental findings at a neuronal, microscopic level. Here, we show that our neuronal-computational understanding of speech learning and recognition may be vastly improved by looking at an animal model, i.e., the songbird, which faces the same challenge as humans: to learn and decode complex auditory input, in an online fashion. Motivated by striking similarities between the human and songbird neural recognition systems at the macroscopic level, we assumed that the human brain uses the same computational principles at a microscopic level and translated a birdsong model into a novel human sound learning and recognition model with an emphasis on speech. We show that the resulting Bayesian model with a hierarchy of nonlinear dynamical systems can learn speech samples such as words rapidly and recognize them robustly, even in adverse conditions. In addition, we show that recognition can be performed even when words are spoken by different speakers and with different accents—an everyday situation in which current state-of-the-art speech recognition models often fail. The model can also be used to qualitatively explain behavioral data on human speech learning and derive predictions for future experiments. PMID:24068902

  13. From birdsong to human speech recognition: bayesian inference on a hierarchy of nonlinear dynamical systems.

    PubMed

    Yildiz, Izzet B; von Kriegstein, Katharina; Kiebel, Stefan J

    2013-01-01

    Our knowledge about the computational mechanisms underlying human learning and recognition of sound sequences, especially speech, is still very limited. One difficulty in deciphering the exact means by which humans recognize speech is that there are scarce experimental findings at a neuronal, microscopic level. Here, we show that our neuronal-computational understanding of speech learning and recognition may be vastly improved by looking at an animal model, i.e., the songbird, which faces the same challenge as humans: to learn and decode complex auditory input, in an online fashion. Motivated by striking similarities between the human and songbird neural recognition systems at the macroscopic level, we assumed that the human brain uses the same computational principles at a microscopic level and translated a birdsong model into a novel human sound learning and recognition model with an emphasis on speech. We show that the resulting Bayesian model with a hierarchy of nonlinear dynamical systems can learn speech samples such as words rapidly and recognize them robustly, even in adverse conditions. In addition, we show that recognition can be performed even when words are spoken by different speakers and with different accents-an everyday situation in which current state-of-the-art speech recognition models often fail. The model can also be used to qualitatively explain behavioral data on human speech learning and derive predictions for future experiments.

  14. Exploring expressivity and emotion with artificial voice and speech technologies.

    PubMed

    Pauletto, Sandra; Balentine, Bruce; Pidcock, Chris; Jones, Kevin; Bottaci, Leonardo; Aretoulaki, Maria; Wells, Jez; Mundy, Darren P; Balentine, James

    2013-10-01

    Emotion in audio-voice signals, as synthesized by text-to-speech (TTS) technologies, was investigated to formulate a theory of expression for user interface design. Emotional parameters were specified with markup tags, and the resulting audio was further modulated with post-processing techniques. Software was then developed to link a selected TTS synthesizer with an automatic speech recognition (ASR) engine, producing a chatbot that could speak and listen. Using these two artificial voice subsystems, investigators explored both artistic and psychological implications of artificial speech emotion. Goals of the investigation were interdisciplinary, with interest in musical composition, augmentative and alternative communication (AAC), commercial voice announcement applications, human-computer interaction (HCI), and artificial intelligence (AI). The work-in-progress points towards an emerging interdisciplinary ontology for artificial voices. As one study output, HCI tools are proposed for future collaboration.

  15. Therapist and Client Interactions in Motivational Interviewing for Social Anxiety Disorder.

    PubMed

    Romano, Mia; Arambasic, Jelena; Peters, Lorna

    2017-07-01

    The aim of the present study is to assess the bidirectional associations between therapist and client speech during a treatment based on motivational interviewing (MI) for social anxiety disorder. Participants were 85 adults diagnosed with social anxiety who received MI prior to entering cognitive behavioral therapy. MI sessions were sequentially coded using the Motivational Interviewing Skill Code 2.5. Therapist MI-consistent behaviors, including open questions as well as positive and negative reflections, were more likely to be followed by client change exploration (change talk and counter-change talk). Therapist MI-inconsistent behaviors were more likely to precede client neutral language. Client language was also found to influence therapist likelihood of responding in an MI-consistent manner. The findings support the first step of the MI causal model in the context of social anxiety and direct future research into the effect of therapist and client behaviors on MI treatment outcome. © 2016 Wiley Periodicals, Inc.

  16. Unilateral Hearing Loss: Understanding Speech Recognition and Localization Variability-Implications for Cochlear Implant Candidacy.

    PubMed

    Firszt, Jill B; Reeder, Ruth M; Holden, Laura K

    At a minimum, unilateral hearing loss (UHL) impairs sound localization ability and understanding speech in noisy environments, particularly if the loss is severe to profound. Accompanying the numerous negative consequences of UHL is considerable unexplained individual variability in the magnitude of its effects. Identification of covariables that affect outcome and contribute to variability in UHLs could augment counseling, treatment options, and rehabilitation. Cochlear implantation as a treatment for UHL is on the rise yet little is known about factors that could impact performance or whether there is a group at risk for poor cochlear implant outcomes when hearing is near-normal in one ear. The overall goal of our research is to investigate the range and source of variability in speech recognition in noise and localization among individuals with severe to profound UHL and thereby help determine factors relevant to decisions regarding cochlear implantation in this population. The present study evaluated adults with severe to profound UHL and adults with bilateral normal hearing. Measures included adaptive sentence understanding in diffuse restaurant noise, localization, roving-source speech recognition (words from 1 of 15 speakers in a 140° arc), and an adaptive speech-reception threshold psychoacoustic task with varied noise types and noise-source locations. There were three age-sex-matched groups: UHL (severe to profound hearing loss in one ear and normal hearing in the contralateral ear), normal hearing listening bilaterally, and normal hearing listening unilaterally. Although the normal-hearing-bilateral group scored significantly better and had less performance variability than UHLs on all measures, some UHL participants scored within the range of the normal-hearing-bilateral group on all measures. The normal-hearing participants listening unilaterally had better monosyllabic word understanding than UHLs for words presented on the blocked/deaf side but not the open/hearing side. In contrast, UHLs localized better than the normal-hearing unilateral listeners for stimuli on the open/hearing side but not the blocked/deaf side. This suggests that UHLs had learned strategies for improved localization on the side of the intact ear. The UHL and unilateral normal-hearing participant groups were not significantly different for speech in noise measures. UHL participants with childhood rather than recent hearing loss onset localized significantly better; however, these two groups did not differ for speech recognition in noise. Age at onset in UHL adults appears to affect localization ability differently than understanding speech in noise. Hearing thresholds were significantly correlated with speech recognition for UHL participants but not the other two groups. Auditory abilities of UHLs varied widely and could be explained only in part by hearing threshold levels. Age at onset and length of hearing loss influenced performance on some, but not all measures. Results support the need for a revised and diverse set of clinical measures, including sound localization, understanding speech in varied environments, and careful consideration of functional abilities as individuals with severe to profound UHL are being considered potential cochlear implant candidates.

  17. Speech intelligibility in complex acoustic environments in young children

    NASA Astrophysics Data System (ADS)

    Litovsky, Ruth

    2003-04-01

    While the auditory system undergoes tremendous maturation during the first few years of life, it has become clear that in complex scenarios when multiple sounds occur and when echoes are present, children's performance is significantly worse than their adult counterparts. The ability of children (3-7 years of age) to understand speech in a simulated multi-talker environment and to benefit from spatial separation of the target and competing sounds was investigated. In these studies, competing sources vary in number, location, and content (speech, modulated or unmodulated speech-shaped noise and time-reversed speech). The acoustic spaces were also varied in size and amount of reverberation. Finally, children with chronic otitis media who received binaural training were tested pre- and post-training on a subset of conditions. Results indicated the following. (1) Children experienced significantly more masking than adults, even in the simplest conditions tested. (2) When the target and competing sounds were spatially separated speech intelligibility improved, but the amount varied with age, type of competing sound, and number of competitors. (3) In a large reverberant classroom there was no benefit of spatial separation. (4) Binaural training improved speech intelligibility performance in children with otitis media. Future work includes similar studies in children with unilateral and bilateral cochlear implants. [Work supported by NIDCD, DRF, and NOHR.

  18. Effect of classroom acoustics on the speech intelligibility of students.

    PubMed

    Rabelo, Alessandra Terra Vasconcelos; Santos, Juliana Nunes; Oliveira, Rafaella Cristina; Magalhães, Max de Castro

    2014-01-01

    To analyze the acoustic parameters of classrooms and the relationship among equivalent sound pressure level (Leq), reverberation time (T₃₀), the Speech Transmission Index (STI), and the performance of students in speech intelligibility testing. A cross-sectional descriptive study, which analyzed the acoustic performance of 18 classrooms in 9 public schools in Belo Horizonte, Minas Gerais, Brazil, was conducted. The following acoustic parameters were measured: Leq, T₃₀, and the STI. In the schools evaluated, a speech intelligibility test was performed on 273 students, 45.4% of whom were boys, with an average age of 9.4 years. The results of the speech intelligibility test were compared to the values of the acoustic parameters with the help of Student's t-test. The Leq, T₃₀, and STI tests were conducted in empty and furnished classrooms. Children showed better results in speech intelligibility tests conducted in classrooms with less noise, a lower T₃₀, and greater STI values. The majority of classrooms did not meet the recommended regulatory standards for good acoustic performance. Acoustic parameters have a direct effect on the speech intelligibility of students. Noise contributes to a decrease in their understanding of information presented orally, which can lead to negative consequences in their education and their social integration as future professionals.

  19. Translations on Eastern Europe, Political, Sociological, and Military Affairs, Number 1413.

    DTIC Science & Technology

    1977-07-06

    Chankova Kalincheva Kiril Georgiev Kotsaliev Konstantin Antonov Kutev Kostadin Dimitrov Ivanov Kaiina Stefanova Koleva Krasimir Nyagolov Nedelchev...10 May 77) 21 Members of Dimitrov Youth Union Leading Bodies (NARODNA MLADEZH, 12 May 77) 79 Concluding Speech of Youth Union Leader... Dimitrov Communist Youth Union opened its proceedings on 9 May at 8:30 am at the Universiada Hall, in Sofia. The congress was opened by Boycho

  20. Asynchronous glimpsing of speech: Spread of masking and task set-size

    PubMed Central

    Ozmeral, Erol J.; Buss, Emily; Hall, Joseph W.

    2012-01-01

    Howard-Jones and Rosen [(1993). J. Acoust. Soc. Am. 93, 2915–2922] investigated the ability to integrate glimpses of speech that are separated in time and frequency using a “checkerboard” masker, with asynchronous amplitude modulation (AM) across frequency. Asynchronous glimpsing was demonstrated only for spectrally wide frequency bands. It is possible that the reduced evidence of spectro-temporal integration with narrower bands was due to spread of masking at the periphery. The present study tested this hypothesis with a dichotic condition, in which the even- and odd-numbered bands of the target speech and asynchronous AM masker were presented to opposite ears, minimizing the deleterious effects of masking spread. For closed-set consonant recognition, thresholds were 5.1–8.5 dB better for dichotic than for monotic asynchronous AM conditions. Results were similar for closed-set word recognition, but for open-set word recognition the benefit of dichotic presentation was more modest and level dependent, consistent with the effects of spread of masking being level dependent. There was greater evidence of asynchronous glimpsing in the open-set than closed-set tasks. Presenting stimuli dichotically supported asynchronous glimpsing with narrower frequency bands than previously shown, though the magnitude of glimpsing was reduced for narrower bandwidths even in some dichotic conditions. PMID:22894234

  1. The speech focus position effect on jaw-finger coordination in a pointing task.

    PubMed

    Rochet-Capellan, Amélie; Laboissière, Rafael; Galván, Arturo; Schwartz, Jean-Luc

    2008-12-01

    This article investigates jaw-finger coordination in a task involving pointing to a target while naming it with a CVCV (e.g., /papa/) versus CVCV (e.g., /papa/) word. According to the authors' working hypothesis, the pointing apex (gesture extremum) would be synchronized with the apex of the jaw-opening gesture corresponding to the stressed syllable. Jaw and finger motions were recorded using Optotrak (Northern Digital, Waterloo, Ontario, Canada). The effects of stress position on jaw-finger coordination were tested across different target positions (near vs. far) and different consonants in the target word (/t/ vs. /p/). Twenty native Portuguese Brazilian speakers participated in the experiment (all conditions). Jaw response starts earlier, and finger-target alignment period is longer for CVCV words than for CVCV ones. The apex of the jaw-opening gesture for the stressed syllable appears synchronized with the onset of the finger-target alignment period (corresponding to the pointing apex) for CVCV words and with the offset of that period for CVCV words. For both stress conditions, the stressed syllable occurs within the finger-target alignment period because of tight finger-jaw coordination. This result is interpreted as evidence for an anchoring of the speech deictic site (part of speech that shows) in the pointing gesture.

  2. Investigation of in-vehicle speech intelligibility metrics for normal hearing and hearing impaired listeners

    NASA Astrophysics Data System (ADS)

    Samardzic, Nikolina

    The effectiveness of in-vehicle speech communication can be a good indicator of the perception of the overall vehicle quality and customer satisfaction. Currently available speech intelligibility metrics do not account in their procedures for essential parameters needed for a complete and accurate evaluation of in-vehicle speech intelligibility. These include the directivity and the distance of the talker with respect to the listener, binaural listening, hearing profile of the listener, vocal effort, and multisensory hearing. In the first part of this research the effectiveness of in-vehicle application of these metrics is investigated in a series of studies to reveal their shortcomings, including a wide range of scores resulting from each of the metrics for a given measurement configuration and vehicle operating condition. In addition, the nature of a possible correlation between the scores obtained from each metric is unknown. The metrics and the subjective perception of speech intelligibility using, for example, the same speech material have not been compared in literature. As a result, in the second part of this research, an alternative method for speech intelligibility evaluation is proposed for use in the automotive industry by utilizing a virtual reality driving environment for ultimately setting targets, including the associated statistical variability, for future in-vehicle speech intelligibility evaluation. The Speech Intelligibility Index (SII) was evaluated at the sentence Speech Receptions Threshold (sSRT) for various listening situations and hearing profiles using acoustic perception jury testing and a variety of talker and listener configurations and background noise. In addition, the effect of individual sources and transfer paths of sound in an operating vehicle to the vehicle interior sound, specifically their effect on speech intelligibility was quantified, in the framework of the newly developed speech intelligibility evaluation method. Lastly, as an example of the significance of speech intelligibility evaluation in the context of an applicable listening environment, as indicated in this research, it was found that the jury test participants required on average an approximate 3 dB increase in sound pressure level of speech material while driving and listening compared to when just listening, for an equivalent speech intelligibility performance and the same listening task.

  3. Students in Action.

    ERIC Educational Resources Information Center

    Insights on Law & Society, 2001

    2001-01-01

    Focuses on important issues surrounding the opening clauses of the First Amendment on the establishment of religion, freedom of speech, and freedom of religion. Includes articles: "Established Churches in Colonial Times,""Establishment Clause: Here's How to Understand it!", "Religious Freedom and Today's Religious…

  4. Simulating reading acquisition: The link between reading outcome and multimodal brain signatures of letter-speech sound learning in prereaders.

    PubMed

    Karipidis, Iliana I; Pleisch, Georgette; Brandeis, Daniel; Roth, Alexander; Röthlisberger, Martina; Schneebeli, Maya; Walitza, Susanne; Brem, Silvia

    2018-05-08

    During reading acquisition, neural reorganization of the human brain facilitates the integration of letters and speech sounds, which enables successful reading. Neuroimaging and behavioural studies have established that impaired audiovisual integration of letters and speech sounds is a core deficit in individuals with developmental dyslexia. This longitudinal study aimed to identify neural and behavioural markers of audiovisual integration that are related to future reading fluency. We simulated the first step of reading acquisition by performing artificial-letter training with prereading children at risk for dyslexia. Multiple logistic regressions revealed that our training provides new precursors of reading fluency at the beginning of reading acquisition. In addition, an event-related potential around 400 ms and functional magnetic resonance imaging activation patterns in the left planum temporale to audiovisual correspondences improved cross-validated prediction of future poor readers. Finally, an exploratory analysis combining simultaneously acquired electroencephalography and hemodynamic data suggested that modulation of temporoparietal brain regions depended on future reading skills. The multimodal approach demonstrates neural adaptations to audiovisual integration in the developing brain that are related to reading outcome. Despite potential limitations arising from the restricted sample size, our results may have promising implications both for identifying poor-reading children and for monitoring early interventions.

  5. Single-Sided Deafness: Impact of Cochlear Implantation on Speech Perception in Complex Noise and on Auditory Localization Accuracy.

    PubMed

    Döge, Julia; Baumann, Uwe; Weissgerber, Tobias; Rader, Tobias

    2017-12-01

    To assess auditory localization accuracy and speech reception threshold (SRT) in complex noise conditions in adult patients with acquired single-sided deafness, after intervention with a cochlear implant (CI) in the deaf ear. Nonrandomized, open, prospective patient series. Tertiary referral university hospital. Eleven patients with late-onset single-sided deafness (SSD) and normal hearing in the unaffected ear, who received a CI. All patients were experienced CI users. Unilateral cochlear implantation. Speech perception was tested in a complex multitalker equivalent noise field consisting of multiple sound sources. Speech reception thresholds in noise were determined in aided (with CI) and unaided conditions. Localization accuracy was assessed in complete darkness. Acoustic stimuli were radiated by multiple loudspeakers distributed in the frontal horizontal plane between -60 and +60 degrees. In the aided condition, results show slightly improved speech reception scores compared with the unaided condition in most of the patients. For 8 of the 11 subjects, SRT was improved between 0.37 and 1.70 dB. Three of the 11 subjects showed deteriorations between 1.22 and 3.24 dB SRT. Median localization error decreased significantly by 12.9 degrees compared with the unaided condition. CI in single-sided deafness is an effective treatment to improve the auditory localization accuracy. Speech reception in complex noise conditions is improved to a lesser extent in 73% of the participating CI SSD patients. However, the absence of true binaural interaction effects (summation, squelch) impedes further improvements. The development of speech processing strategies that respect binaural interaction seems to be mandatory to advance speech perception in demanding listening situations in SSD patients.

  6. Late onset canonical babbling: a possible early marker of abnormal development.

    PubMed

    Oller, D K; Eilers, R E; Neal, A R; Cobo-Lewis, A B

    1998-11-01

    By their 10th month of life, typically developing infants produce canonical babbling, which includes the well-formed syllables required for meaningful speech. Research suggests that emerging speech or language-related disorders might be associated with late onset of canonical babbling. Onset of canonical babbling was investigated for 1,536 high-risk infants, at about 10-months corrected age. Parental report by open-ended questionnaire was found to be an efficient method for ascertaining babbling status. Although delays were infrequent, they were often associated with genetic, neurological, anatomical, and/or physiological abnormalities. Over half the cases of late canonical babbling were not, at the time they were discovered associated with prior significant medical diagnoses. Late canonical-babbling onset may be a predictor of later developmental disabilities, including problems in speech, language, and reading.

  7. Stuttering treatment for a school-age child with Down syndrome: a descriptive case report.

    PubMed

    Harasym, Jessica; Langevin, Marilyn

    2012-12-01

    Little is known about optimal treatment approaches and stuttering treatment outcomes for children with Down syndrome. The purpose of this study was to investigate outcomes for a child with Down syndrome who received a combination of fluency shaping therapy and parent delivered contingencies for normally fluent speech, prolonged speech, and stuttered speech. In-clinic speech measures obtained at post-treatment and at 4 months follow-up reflected improvements in fluency of 89.0% and 98.6%, respectively. The participant's beyond-clinic follow-up sample reflected an improvement of 95.5%. Following treatment, the participant demonstrated improved self-confidence, self-esteem, and improved participation and functioning at school. Findings suggest that fluency shaping with parental contingencies may be a viable treatment approach to reduce stuttering in children with Down syndrome. Future research using an experimental research design is warranted. Readers will be able to describe (a) prevalence estimates of stuttering in individuals with Down syndrome, (b) the main components of a fluency shaping program for a child with Down syndrome who stutters and has co-occurring speech and language delays, and (c) speech and parent-, teacher-, and self-report treatment outcomes. Copyright © 2012 Elsevier Inc. All rights reserved.

  8. Automatic detection of obstructive sleep apnea using speech signals.

    PubMed

    Goldshtein, Evgenia; Tarasiuk, Ariel; Zigel, Yaniv

    2011-05-01

    Obstructive sleep apnea (OSA) is a common disorder associated with anatomical abnormalities of the upper airways that affects 5% of the population. Acoustic parameters may be influenced by the vocal tract structure and soft tissue properties. We hypothesize that speech signal properties of OSA patients will be different than those of control subjects not having OSA. Using speech signal processing techniques, we explored acoustic speech features of 93 subjects who were recorded using a text-dependent speech protocol and a digital audio recorder immediately prior to polysomnography study. Following analysis of the study, subjects were divided into OSA (n=67) and non-OSA (n=26) groups. A Gaussian mixture model-based system was developed to model and classify between the groups; discriminative features such as vocal tract length and linear prediction coefficients were selected using feature selection technique. Specificity and sensitivity of 83% and 79% were achieved for the male OSA and 86% and 84% for the female OSA patients, respectively. We conclude that acoustic features from speech signals during wakefulness can detect OSA patients with good specificity and sensitivity. Such a system can be used as a basis for future development of a tool for OSA screening. © 2011 IEEE

  9. Real-time classification of auditory sentences using evoked cortical activity in humans

    NASA Astrophysics Data System (ADS)

    Moses, David A.; Leonard, Matthew K.; Chang, Edward F.

    2018-06-01

    Objective. Recent research has characterized the anatomical and functional basis of speech perception in the human auditory cortex. These advances have made it possible to decode speech information from activity in brain regions like the superior temporal gyrus, but no published work has demonstrated this ability in real-time, which is necessary for neuroprosthetic brain-computer interfaces. Approach. Here, we introduce a real-time neural speech recognition (rtNSR) software package, which was used to classify spoken input from high-resolution electrocorticography signals in real-time. We tested the system with two human subjects implanted with electrode arrays over the lateral brain surface. Subjects listened to multiple repetitions of ten sentences, and rtNSR classified what was heard in real-time from neural activity patterns using direct sentence-level and HMM-based phoneme-level classification schemes. Main results. We observed single-trial sentence classification accuracies of 90% or higher for each subject with less than 7 minutes of training data, demonstrating the ability of rtNSR to use cortical recordings to perform accurate real-time speech decoding in a limited vocabulary setting. Significance. Further development and testing of the package with different speech paradigms could influence the design of future speech neuroprosthetic applications.

  10. Assessment of Individuals with Primary Progressive Aphasia.

    PubMed

    Henry, Maya L; Grasso, Stephanie M

    2018-07-01

    Speech-language pathologists play a crucial role in the assessment and treatment of individuals with primary progressive aphasia (PPA). The speech-language evaluation is a critical aspect of the diagnostic and rehabilitative process, informing differential diagnosis as well as intervention planning and monitoring of cognitive-linguistic status over time. The evaluation should include a thorough case history and interview and a detailed assessment of speech-language and cognitive functions, with tasks designed to detect core and associated deficits outlined in current diagnostic criteria. In this paper, we review assessments that can be utilized to examine communication and cognition in PPA, including general aphasia batteries designed for stroke and/or progressive aphasia as well as tests of specific cognitive-linguistic functions, including naming, object/person knowledge, single-word and sentence comprehension, repetition, spontaneous speech/language production, motor speech, written language, and nonlinguistic cognitive domains. The comprehensive evaluation can inform diagnostic decision making and facilitate planning of interventions that are tailored to the patient's current status and likely progression of deficits. As such, the speech-language evaluation allows the medical team to provide individuals with PPA and their families with appropriate recommendations for the present and the future. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  11. Speech impairment in Down syndrome: a review.

    PubMed

    Kent, Ray D; Vorperian, Houri K

    2013-02-01

    This review summarizes research on disorders of speech production in Down syndrome (DS) for the purposes of informing clinical services and guiding future research. Review of the literature was based on searches using MEDLINE, Google Scholar, PsycINFO, and HighWire Press, as well as consideration of reference lists in retrieved documents (including online sources). Search terms emphasized functions related to voice, articulation, phonology, prosody, fluency, and intelligibility. The following conclusions pertain to four major areas of review: voice, speech sounds, fluency and prosody, and intelligibility. The first major area is voice. Although a number of studies have reported on vocal abnormalities in DS, major questions remain about the nature and frequency of the phonatory disorder. Results of perceptual and acoustic studies have been mixed, making it difficult to draw firm conclusions or even to identify sensitive measures for future study. The second major area is speech sounds. Articulatory and phonological studies show that speech patterns in DS are a combination of delayed development and errors not seen in typical development. Delayed (i.e., developmental) and disordered (i.e., nondevelopmental) patterns are evident by the age of about 3 years, although DS-related abnormalities possibly appear earlier, even in infant babbling. The third major area is fluency and prosody. Stuttering and/or cluttering occur in DS at rates of 10%-45%, compared with about 1% in the general population. Research also points to significant disturbances in prosody. The fourth major area is intelligibility. Studies consistently show marked limitations in this area, but only recently has the research gone beyond simple rating scales.

  12. Objective measurement of motor speech characteristics in the healthy pediatric population.

    PubMed

    Wong, A W; Allegro, J; Tirado, Y; Chadha, N; Campisi, P

    2011-12-01

    To obtain objective measurements of motor speech characteristics in normal children, using a computer-based motor speech software program. Cross-sectional, observational design in a university-based ambulatory pediatric otolaryngology clinic. Participants included 112 subjects (54 females and 58 males) aged 4-18 years. Participants with previously diagnosed hearing loss, voice and motor disorders, and children unable to repeat a passage in English were excluded. Voice samples were recorded and analysed using the Motor Speech Profile (MSP) software (KayPENTAX, Lincoln Park, NJ). The MSP produced measures of diadochokinetics, second formant transition, intonation, and syllabic rates. Demographic data, including sex, age, and cigarette smoke exposure were obtained. Normative data for several motor speech characteristics were derived for children ranging from age 4 to 18 years. A number of age-dependent changes were indentified, including an increase in average diadochokinetic rate (p<0.001) and standard syllabic duration (p<0.001) with age. There were no identified differences in motor speech characteristics between males and females across the measured age range. Variations in fundamental frequency (Fo) during speech did not change significantly with age for both males and females. To our knowledge, this is the first pediatric normative database for the MSP progam. The MSP is suitable for testing children and can be used to study developmental changes in motor speech. The analysis demonstrated that males and females behave similarly and show the same relationship with age for the motor speech characteristics studied. This normative database will provide essential comparative data for future studies exploring alterations in motor speech that may occur with hearing, voice, and motor disorders and to assess the results of targeted therapies. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.

  13. JND measurements of the speech formants parameters and its implication in the LPC pole quantization

    NASA Astrophysics Data System (ADS)

    Orgad, Yaakov

    1988-08-01

    The inherent sensitivity of auditory perception is explicitly used with the objective of designing an efficient speech encoder. Speech can be modelled by a filter representing the vocal tract shape that is driven by an excitation signal representing glottal air flow. This work concentrates on the filter encoding problem, assuming that excitation signal encoding is optimal. Linear predictive coding (LPC) techniques were used to model a short speech segment by an all-pole filter; each pole was directly related to the speech formants. Measurements were made of the auditory just noticeable difference (JND) corresponding to the natural speech formants, with the LPC filter poles as the best candidates to represent the speech spectral envelope. The JND is the maximum precision required in speech quantization; it was defined on the basis of the shift of one pole parameter of a single frame of a speech segment, necessary to induce subjective perception of the distortion, with .75 probability. The average JND in LPC filter poles in natural speech was found to increase with increasing pole bandwidth and, to a lesser extent, frequency. The JND measurements showed a large spread of the residuals around the average values, indicating that inter-formant coupling and, perhaps, other, not yet fully understood, factors were not taken into account at this stage of the research. A future treatment should consider these factors. The average JNDs obtained in this work were used to design pole quantization tables for speech coding and provided a better bit-rate than the standard quantizer of reflection coefficient; a 30-bits-per-frame pole quantizer yielded a speech quality similar to that obtained with a standard 41-bits-per-frame reflection coefficient quantizer. Owing to the complexity of the numerical root extraction system, the practical implementation of the pole quantization approach remains to be proved.

  14. An automatic speech recognition system with speaker-independent identification support

    NASA Astrophysics Data System (ADS)

    Caranica, Alexandru; Burileanu, Corneliu

    2015-02-01

    The novelty of this work relies on the application of an open source research software toolkit (CMU Sphinx) to train, build and evaluate a speech recognition system, with speaker-independent support, for voice-controlled hardware applications. Moreover, we propose to use the trained acoustic model to successfully decode offline voice commands on embedded hardware, such as an ARMv6 low-cost SoC, Raspberry PI. This type of single-board computer, mainly used for educational and research activities, can serve as a proof-of-concept software and hardware stack for low cost voice automation systems.

  15. Quantitative assessment of motor speech abnormalities in idiopathic rapid eye movement sleep behaviour disorder.

    PubMed

    Rusz, Jan; Hlavnička, Jan; Tykalová, Tereza; Bušková, Jitka; Ulmanová, Olga; Růžička, Evžen; Šonka, Karel

    2016-03-01

    Patients with idiopathic rapid eye movement sleep behaviour disorder (RBD) are at substantial risk for developing Parkinson's disease (PD) or related neurodegenerative disorders. Speech is an important indicator of motor function and movement coordination, and therefore may be an extremely sensitive early marker of changes due to prodromal neurodegeneration. Speech data were acquired from 16 RBD subjects and 16 age- and sex-matched healthy control subjects. Objective acoustic assessment of 15 speech dimensions representing various phonatory, articulatory, and prosodic deviations was performed. Statistical models were applied to characterise speech disorders in RBD and to estimate sensitivity and specificity in differentiating between RBD and control subjects. Some form of speech impairment was revealed in 88% of RBD subjects. Articulatory deficits were the most prominent findings in RBD. In comparison to controls, the RBD group showed significant alterations in irregular alternating motion rates (p = 0.009) and articulatory decay (p = 0.01). The combination of four distinctive speech dimensions, including aperiodicity, irregular alternating motion rates, articulatory decay, and dysfluency, led to 96% sensitivity and 79% specificity in discriminating between RBD and control subjects. Speech impairment was significantly more pronounced in RBD subjects with the motor score of the Unified Parkinson's Disease Rating Scale greater than 4 points when compared to other RBD individuals. Simple quantitative speech motor measures may be suitable for the reliable detection of prodromal neurodegeneration in subjects with RBD, and therefore may provide important outcomes for future therapy trials. Copyright © 2015 Elsevier B.V. All rights reserved.

  16. Association between speech-language, general cognitive functioning and behaviour problems in individuals with Williams syndrome.

    PubMed

    Rossi, N F; Giacheti, C M

    2017-07-01

    Williams syndrome (WS) phenotype is described as unique and intriguing. The aim of this study was to investigate the associations between speech-language abilities, general cognitive functioning and behavioural problems in individuals with WS, considering age effects and speech-language characteristics of WS sub-groups. The study's participants were 26 individuals with WS and their parents. General cognitive functioning was assessed with the Wechsler Intelligence Scale. Peabody Picture Vocabulary Test, Token Test and the Cookie Theft Picture test were used as speech-language measures. Five speech-language characteristics were evaluated from a 30-min conversation (clichés, echolalia, perseverative speech, exaggerated prosody and monotone intonation). The Child Behaviour Checklist (CBCL 6-18) was used to assess behavioural problems. Higher single-word receptive vocabulary and narrative vocabulary were negatively associated with CBCL T-scores for Social Problems, Aggressive Behaviour and Total Problems. Speech rate was negatively associated with the CBCL Withdrawn/Depressed T-score. Monotone intonation was associated with shy behaviour, as well as exaggerated prosody with talkative behaviour. WS with perseverative speech and exaggerated prosody presented higher scores on Thought Problems. Echolalia was significantly associated with lower Verbal IQ. No significant association was found between IQ and behaviour problems. Age-associated effects were observed only for the Aggressive Behaviour scale. Associations reported in the present study may represent an insightful background for future predictive studies of speech-language, cognition and behaviour problems in WS. © 2017 MENCAP and International Association of the Scientific Study of Intellectual and Developmental Disabilities and John Wiley & Sons Ltd.

  17. Perception and analysis of Spanish accents in English speech

    NASA Astrophysics Data System (ADS)

    Chism, Cori; Lass, Norman

    2002-05-01

    The purpose of the present study was to determine what relates most closely to the degree of perceived foreign accent in the English speech of native Spanish speakers: intonation, vowel length, stress, voice onset time (VOT), or segmental accuracy. Nineteen native English speaking listeners rated speech samples from 7 native English speakers and 15 native Spanish speakers for comprehensibility and degree of foreign accent. The speech samples were analyzed spectrographically and perceptually to obtain numerical values for each variable. Correlation coefficients were computed to determine the relationship beween these values and the average foreign accent scores. Results showed that the average foreign accent scores were statistically significantly correlated with three variables: the length of stressed vowels (r=-0.48, p=0.05), voice onset time (r =-0.62, p=0.01), and segmental accuracy (r=0.92, p=0.001). Implications of these findings and suggestions for future research are discussed.

  18. Five analogies between a King's Speech treatment and contemporary play therapies.

    PubMed

    Terr, Lenore C

    2012-01-01

    Psychiatric patients frequently respond positively to play therapy, which may rely on psychoanalytic, Jungian, cognitive-behavioral, familial, school-based, or other theories. I wished to determine if there were unifying principles that tie together these various types of play treatments. The fact-based film, The King's Speech, vividly illustrates play utilized by Lionel Logue in his speech treatment (1926-1939) of the future King of England. In the film I found five analogies to the play therapy I employ in office practice. The play scenes in The King's Speech point to five unifying principles among contemporary play therapies: (1) the crucial nature of the relationship, (2) the centrality of having fun, (3) the occasional reliance on others, (4) the interjection of pithy talk, and (5) the usefulness of a little drama. No matter what theory a play therapist ascribes to, these five unifying principles should be kept in mind during treatment.

  19. Construct-related validity of the TOCS measures: comparison of intelligibility and speaking rate scores in children with and without speech disorders.

    PubMed

    Hodge, Megan M; Gotzke, Carrie L

    2014-01-01

    This study evaluated construct-related validity of the Test of Children's Speech (TOCS). Intelligibility scores obtained using open-set word identification tasks (orthographic transcription) for the TOCS word and sentence tests and rate scores for the TOCS sentence test (words per minute or WPM and intelligible words per minute or IWPM) were compared for a group of 15 adults (18-30 years of age) with normal speech production and three groups of children: 48 3-6 year-olds with typical speech development and neurological histories (TDS), 48 3-6 year-olds with a speech sound disorder of unknown origin and no identified neurological impairment (SSD-UNK), and 22 3-10 year-olds with dysarthria and cerebral palsy (DYS). As expected, mean intelligibility scores and rates increased with age in the TDS group. However, word test intelligibility, WPM and IWPM scores for the 6 year-olds in the TDS group were significantly lower than those for the adults. The DYS group had significantly lower word and sentence test intelligibility and WPM and IWPM scores than the TDS and SSD-UNK groups. Compared to the TDS group, the SSD-UNK group also had significantly lower intelligibility scores for the word and sentence tests, and significantly lower IWPM, but not WPM scores on the sentence test. The results support the construct-related validity of TOCS as a tool for obtaining intelligibility and rate scores that are sensitive to group differences in 3-6 year-old children, with and without speech sound disorders, and to 3+ year-old children with speech disorders, with and without dysarthria. Readers will describe the word and sentence intelligibility and speaking rate performance of children with typically developing speech at age levels of 3, 4, 5 and 6 years, as measured by the Test of Children's Speech, and how these compare with adult speakers and two groups of children with speech disorders. They will also recognize what measures on this test differentiate children with speech sound disorders of unknown origin from children with cerebral palsy and dysarthria. Copyright © 2014 Elsevier Inc. All rights reserved.

  20. Shared acoustic codes underlie emotional communication in music and speech—Evidence from deep transfer learning

    PubMed Central

    Schuller, Björn

    2017-01-01

    Music and speech exhibit striking similarities in the communication of emotions in the acoustic domain, in such a way that the communication of specific emotions is achieved, at least to a certain extent, by means of shared acoustic patterns. From an Affective Sciences points of view, determining the degree of overlap between both domains is fundamental to understand the shared mechanisms underlying such phenomenon. From a Machine learning perspective, the overlap between acoustic codes for emotional expression in music and speech opens new possibilities to enlarge the amount of data available to develop music and speech emotion recognition systems. In this article, we investigate time-continuous predictions of emotion (Arousal and Valence) in music and speech, and the Transfer Learning between these domains. We establish a comparative framework including intra- (i.e., models trained and tested on the same modality, either music or speech) and cross-domain experiments (i.e., models trained in one modality and tested on the other). In the cross-domain context, we evaluated two strategies—the direct transfer between domains, and the contribution of Transfer Learning techniques (feature-representation-transfer based on Denoising Auto Encoders) for reducing the gap in the feature space distributions. Our results demonstrate an excellent cross-domain generalisation performance with and without feature representation transfer in both directions. In the case of music, cross-domain approaches outperformed intra-domain models for Valence estimation, whereas for Speech intra-domain models achieve the best performance. This is the first demonstration of shared acoustic codes for emotional expression in music and speech in the time-continuous domain. PMID:28658285

  1. Effects of Long-Term Speech-in-Noise Training in Air Traffic Controllers and High Frequency Suppression. A Control Group Study.

    PubMed

    Pérez Zaballos, María Teresa; Ramos de Miguel, Ángel; Pérez Plasencia, Daniel; Zaballos González, María Luisa; Ramos Macías, Ángel

    2015-12-01

    To evaluate 1) if air traffic controllers (ATC) perform better than non-air traffic controllers in an open-set speech-in-noise test because of their experience with radio communications, and 2) if high-frequency information (>8000 Hz) substantially improves speech-in-noise perception across populations. The control group comprised 28 normal-hearing subjects, and the target group comprised 48 ATCs aged between 19 and 55 years who were native Spanish speakers. The hearing -in-noise abilities of the two groups were characterized under two signal conditions: 1) speech tokens and white noise sampled at 44.1 kHz (unfiltered condition) and 2) speech tokens plus white noise, each passed through a 4th order Butterworth filter with 70 and 8000 Hz low and high cutoffs (filtered condition). These tests were performed at signal-to-noise ratios of +5, 0, and -5-dB SNR. The ATCs outperformed the control group in all conditions. The differences were statistically significant in all cases, and the largest difference was observed under the most difficult conditions (-5 dB SNR). Overall, scores were higher when high-frequency components were not suppressed for both groups, although statistically significant differences were not observed for the control group at 0 dB SNR. The results indicate that ATCs are more capable of identifying speech in noise. This may be due to the effect of their training. On the other hand, performance seems to decrease when the high frequency components of speech are removed, regardless of training.

  2. Systematic studies of modified vocalization: speech production changes during a variation of metronomic speech in persons who do and do not stutter.

    PubMed

    Davidow, Jason H; Bothe, Anne K; Ye, Jun

    2011-06-01

    The most common way to induce fluency using rhythm requires persons who stutter to speak one syllable or one word to each beat of a metronome, but stuttering can also be eliminated when the stimulus is of a particular duration (e.g., 1 second [s]). The present study examined stuttering frequency, speech production changes, and speech naturalness during rhythmic speech that alternated 1s of reading with 1s of silence. A repeated-measures design was used to compare data obtained during a control reading condition and during rhythmic reading in 10 persons who stutter (PWS) and 10 normally fluent controls. Ratings for speech naturalness were also gathered from naïve listeners. Results showed that mean vowel duration increased significantly, and the percentage of short phonated intervals decreased significantly, for both groups from the control to the experimental condition. Mean phonated interval length increased significantly for the fluent controls. Mean speech naturalness ratings during the experimental condition were approximately "7" on a 1-9 scale (1=highly natural; 9=highly unnatural), and these ratings were significantly correlated with vowel duration and phonated intervals for PWS. The findings indicate that PWS may be altering vocal fold vibration duration to obtain fluency during this rhythmic speech style, and that vocal fold vibration duration may have an impact on speech naturalness during rhythmic speech. Future investigations should examine speech production changes and speech naturalness during variations of this rhythmic condition. The reader will be able to: (1) describe changes (from a control reading condition) in speech production variables when alternating between 1s of reading and 1s of silence, (2) describe which rhythmic conditions have been found to sound and feel the most natural, (3) describe methodological issues for studies about alterations in speech production variables during fluency-inducing conditions, and (4) describe which fluency-inducing conditions have been shown to involve a reduction in short phonated intervals. Copyright © 2011 Elsevier Inc. All rights reserved.

  3. Speech perception in autism spectrum disorder: An activation likelihood estimation meta-analysis.

    PubMed

    Tryfon, Ana; Foster, Nicholas E V; Sharda, Megha; Hyde, Krista L

    2018-02-15

    Autism spectrum disorder (ASD) is often characterized by atypical language profiles and auditory and speech processing. These can contribute to aberrant language and social communication skills in ASD. The study of the neural basis of speech perception in ASD can serve as a potential neurobiological marker of ASD early on, but mixed results across studies renders it difficult to find a reliable neural characterization of speech processing in ASD. To this aim, the present study examined the functional neural basis of speech perception in ASD versus typical development (TD) using an activation likelihood estimation (ALE) meta-analysis of 18 qualifying studies. The present study included separate analyses for TD and ASD, which allowed us to examine patterns of within-group brain activation as well as both common and distinct patterns of brain activation across the ASD and TD groups. Overall, ASD and TD showed mostly common brain activation of speech processing in bilateral superior temporal gyrus (STG) and left inferior frontal gyrus (IFG). However, the results revealed trends for some distinct activation in the TD group showing additional activation in higher-order brain areas including left superior frontal gyrus (SFG), left medial frontal gyrus (MFG), and right IFG. These results provide a more reliable neural characterization of speech processing in ASD relative to previous single neuroimaging studies and motivate future work to investigate how these brain signatures relate to behavioral measures of speech processing in ASD. Copyright © 2017 Elsevier B.V. All rights reserved.

  4. Oral Language and Reading; Proceedings of the Annual Reading Conference of the Department of Elementary Education at Indiana State University (3rd, Terre Haute, June 14-15, 1973).

    ERIC Educational Resources Information Center

    Waterman, David C., Ed.; Gibbs, Vanita M., Ed.

    This pamphlet is a collection of the speeches given at the Third Annual Reading Conference at Indiana State University, Terre Haute. The theme of the conference was "Oral Language and Reading." The contents include: "Official Program"; opening remarks, "They Led and Followed," by William G. McCarthy; opening address, "Strategies for Reading…

  5. Differences in the speech of 10- to 13-year-old boys from divorced and nondivorced families against the background of emotional attachment.

    PubMed

    Böhm, Birgit

    2004-01-01

    In Germany, an increasing number of children live with one parent alone and have to cope with the separation or divorce of their parents. Emotional drawbacks have frequently been hypothesized for these children. Thus, we studied if such experiences are reflected in speech behavior. Twenty-eight 10- to 13-year-old boys from separated parents (physical separation of the parents was 2 years before the investigation) were compared with 26 boys from parents living together in an interview focusing on attachment-related themes and everyday situations. The interviews were analyzed with regard to coherence of speech, coping with emotional problems, reflectivity, child representation of both parents, and verbal and nonverbal expression of feelings. Boys from separated parents had incoherent speech, difficulties in coping with emotional problems, a poorer reflectivity (thinking about their own mental states and those of others), they represented neither parent supportively and did not show their feelings openly. These results can be traced back to an insecure attachment representation of the boys with separated parents. Copyright 2004 S. Karger AG, Basel

  6. Recreating America's Community Colleges: Implications of the Substantive Issues in Their Future

    ERIC Educational Resources Information Center

    Eaton, Judith S.

    2006-01-01

    CHEA is an institutional membership organization of degree-granting colleges and universities, and serves as an institutional voice for accreditation. CHEA also recognizes 60 regional, national and specialized accrediting organizations. This paper (the text of a speech delivered via teleconference to the Community College Futures Assembly,…

  7. Language for Winning Hearts and Minds: Verb Aspect in U.S. Presidential Campaign Speeches for Engaging Emotion.

    PubMed

    Havas, David A; Chapp, Christopher B

    2016-01-01

    How does language influence the emotions and actions of large audiences? Functionally, emotions help address environmental uncertainty by constraining the body to support adaptive responses and social coordination. We propose emotions provide a similar function in language processing by constraining the mental simulation of language content to facilitate comprehension, and to foster alignment of mental states in message recipients. Consequently, we predicted that emotion-inducing language should be found in speeches specifically designed to create audience alignment - stump speeches of United States presidential candidates. We focused on phrases in the past imperfective verb aspect ("a bad economy was burdening us") that leave a mental simulation of the language content open-ended, and thus unconstrained, relative to past perfective sentences ("we were burdened by a bad economy"). As predicted, imperfective phrases appeared more frequently in stump versus comparison speeches, relative to perfective phrases. In a subsequent experiment, participants rated phrases from presidential speeches as more emotionally intense when written in the imperfective aspect compared to the same phrases written in the perfective aspect, particularly for sentences perceived as negative in valence. These findings are consistent with the notion that emotions have a role in constraining the comprehension of language, a role that may be used in communication with large audiences.

  8. The Future of Software Engineering for High Performance Computing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pope, G

    DOE ASCR requested that from May through mid-July 2015 a study group identify issues and recommend solutions from a software engineering perspective transitioning into the next generation of High Performance Computing. The approach used was to ask some of the DOE complex experts who will be responsible for doing this work to contribute to the study group. The technique used was to solicit elevator speeches: a short and concise write up done as if the author was a speaker with only a few minutes to convince a decision maker of their top issues. Pages 2-18 contain the original texts ofmore » the contributed elevator speeches and end notes identifying the 20 contributors. The study group also ranked the importance of each topic, and those scores are displayed with each topic heading. A perfect score (and highest priority) is three, two is medium priority, and one is lowest priority. The highest scoring topic areas were software engineering and testing resources; the lowest scoring area was compliance to DOE standards. The following two paragraphs are an elevator speech summarizing the contributed elevator speeches. Each sentence or phrase in the summary is hyperlinked to its source via a numeral embedded in the text. A risk one liner has also been added to each topic to allow future risk tracking and mitigation.« less

  9. An examination of speech reception thresholds measured in a simulated reverberant cafeteria environment.

    PubMed

    Best, Virginia; Keidser, Gitte; Buchholz, Jörg M; Freeston, Katrina

    2015-01-01

    There is increasing demand in the hearing research community for the creation of laboratory environments that better simulate challenging real-world listening environments. The hope is that the use of such environments for testing will lead to more meaningful assessments of listening ability, and better predictions about the performance of hearing devices. Here we present one approach for simulating a complex acoustic environment in the laboratory, and investigate the effect of transplanting a speech test into such an environment. Speech reception thresholds were measured in a simulated reverberant cafeteria, and in a more typical anechoic laboratory environment containing background speech babble. The participants were 46 listeners varying in age and hearing levels, including 25 hearing-aid wearers who were tested with and without their hearing aids. Reliable SRTs were obtained in the complex environment, but led to different estimates of performance and hearing-aid benefit from those measured in the standard environment. The findings provide a starting point for future efforts to increase the real-world relevance of laboratory-based speech tests.

  10. An examination of speech reception thresholds measured in a simulated reverberant cafeteria environment

    PubMed Central

    Best, Virginia; Keidser, Gitte; Buchholz, J(x004E7)rg M.; Freeston, Katrina

    2016-01-01

    Objective There is increasing demand in the hearing research community for the creation of laboratory environments that better simulate challenging real-world listening environments. The hope is that the use of such environments for testing will lead to more meaningful assessments of listening ability, and better predictions about the performance of hearing devices. Here we present one approach for simulating a complex acoustic environment in the laboratory, and investigate the effect of transplanting a speech test into such an environment. Design Speech reception thresholds were measured in a simulated reverberant cafeteria, and in a more typical anechoic laboratory environment containing background speech babble. Study Sample The participants were 46 listeners varying in age and hearing levels, including 25 hearing-aid wearers who were tested with and without their hearing aids. Results Reliable SRTs were obtained in the complex environment, but led to different estimates of performance and hearing aid benefit from those measured in the standard environment. Conclusions The findings provide a starting point for future efforts to increase the real-world relevance of laboratory-based speech tests. PMID:25853616

  11. Hemispheric asymmetry in the hierarchical perception of music and speech.

    PubMed

    Rosenthal, Matthew A

    2016-11-01

    The perception of music and speech involves a higher level, cognitive mechanism that allows listeners to form expectations for future music and speech events. This article comprehensively reviews studies on hemispheric differences in the formation of melodic and harmonic expectations in music and selectively reviews studies on hemispheric differences in the formation of syntactic and semantic expectations in speech. On the basis of this review, it is concluded that the higher level mechanism flexibly lateralizes music processing to either hemisphere depending on the expectation generated by a given musical context. When a context generates in the listener an expectation whose elements are sequentially ordered over time, higher level processing is dominant in the left hemisphere. When a context generates in the listener an expectation whose elements are not sequentially ordered over time, higher level processing is dominant in the right hemisphere. This article concludes with a spreading activation model that describes expectations for music and speech in terms of shared temporal and nontemporal representations. (PsycINFO Database Record (c) 2016 APA, all rights reserved).

  12. Military and Government Applications of Human-Machine Communication by Voice

    NASA Astrophysics Data System (ADS)

    Weinstein, Clifford J.

    1995-10-01

    This paper describes a range of opportunities for military and government applications of human-machine communication by voice, based on visits and contacts with numerous user organizations in the United States. The applications include some that appear to be feasible by careful integration of current state-of-the-art technology and others that will require a varying mix of advances in speech technology and in integration of the technology into applications environments. Applications that are described include (1) speech recognition and synthesis for mobile command and control; (2) speech processing for a portable multifunction soldier's computer; (3) speech- and language-based technology for naval combat team tactical training; (4) speech technology for command and control on a carrier flight deck; (5) control of auxiliary systems, and alert and warning generation, in fighter aircraft and helicopters; and (6) voice check-in, report entry, and communication for law enforcement agents or special forces. A phased approach for transfer of the technology into applications is advocated, where integration of applications systems is pursued in parallel with advanced research to meet future needs.

  13. Contrast-marking prosodic emphasis in Williams syndrome: results of detailed phonetic analysis.

    PubMed

    Ito, Kiwako; Martens, Marilee A

    2017-01-01

    Past reports on the speech production of individuals with Williams syndrome (WS) suggest that their prosody is anomalous and may lead to challenges in spoken communication. While existing prosodic assessments confirm that individuals with WS fail to use prosodic emphasis to express contrast, those reports typically lack detailed phonetic analysis of speech data. The present study examines the acoustic properties of speech prosody, aiming for the future development of targeted speech interventions. The study examines the three primary acoustic correlates of prosodic emphasis (duration, intensity, F0) and determines whether individuals with WS have difficulty in producing all or a particular set of the three prosodic cues. Speech produced by 12 individuals with WS and 12 chronological age (CA)-matched typically developing individuals were recorded. A sequential picture-naming task elicited production of target phrases in three contexts: (1) no contrast: gorilla with a racket → rabbit with a balloon; (2) contrast on the animal: fox with a balloon → rabbit with a balloon; and (3) contrast on the object: rabbit with a ball → rabbit with a balloon. The three acoustic correlates of prosodic prominence (duration, intensity and F0) were compared across the three referential contexts. The two groups exhibited striking similarities in their use of word duration and intensity for expressing contrast. Both groups showed the reduction and enhancement of final lengthening, and the enhancement and reduction of intensity difference for the animal contrast and for the object contrast conditions, respectively. The two groups differed in their use of F0: the CA group produced higher F0 for the animal than for the object regardless of the context, and this difference was enhanced when the animal noun was contrastive. In contrast, the WS group produced higher F0 for the object than for the animal when the object was contrastive. The present data contradict previous assessment results that report a lack of prosodic skills to mark contrast in individuals with WS. The methodological differences that may account for this variability are discussed. The present data suggest that individuals with WS produce appropriate prosodic cues to express contrast, although their use of pitch may be somewhat atypical. Additional data and future speech comprehension studies will determine whether pitch modulation can be targeted for speech intervention in individuals with WS. © 2016 Royal College of Speech and Language Therapists.

  14. Initial Progress Toward Development of a Voice-Based Computer-Delivered Motivational Intervention for Heavy Drinking College Students: An Experimental Study

    PubMed Central

    Lechner, William J; MacGlashan, James; Wray, Tyler B; Littman, Michael L

    2017-01-01

    Background Computer-delivered interventions have been shown to be effective in reducing alcohol consumption in heavy drinking college students. However, these computer-delivered interventions rely on mouse, keyboard, or touchscreen responses for interactions between the users and the computer-delivered intervention. The principles of motivational interviewing suggest that in-person interventions may be effective, in part, because they encourage individuals to think through and speak aloud their motivations for changing a health behavior, which current computer-delivered interventions do not allow. Objective The objective of this study was to take the initial steps toward development of a voice-based computer-delivered intervention that can ask open-ended questions and respond appropriately to users’ verbal responses, more closely mirroring a human-delivered motivational intervention. Methods We developed (1) a voice-based computer-delivered intervention that was run by a human controller and that allowed participants to speak their responses to scripted prompts delivered by speech generation software and (2) a text-based computer-delivered intervention that relied on the mouse, keyboard, and computer screen for all interactions. We randomized 60 heavy drinking college students to interact with the voice-based computer-delivered intervention and 30 to interact with the text-based computer-delivered intervention and compared their ratings of the systems as well as their motivation to change drinking and their drinking behavior at 1-month follow-up. Results Participants reported that the voice-based computer-delivered intervention engaged positively with them in the session and delivered content in a manner consistent with motivational interviewing principles. At 1-month follow-up, participants in the voice-based computer-delivered intervention condition reported significant decreases in quantity, frequency, and problems associated with drinking, and increased perceived importance of changing drinking behaviors. In comparison to the text-based computer-delivered intervention condition, those assigned to voice-based computer-delivered intervention reported significantly fewer alcohol-related problems at the 1-month follow-up (incident rate ratio 0.60, 95% CI 0.44-0.83, P=.002). The conditions did not differ significantly on perceived importance of changing drinking or on measures of drinking quantity and frequency of heavy drinking. Conclusions Results indicate that it is feasible to construct a series of open-ended questions and a bank of responses and follow-up prompts that can be used in a future fully automated voice-based computer-delivered intervention that may mirror more closely human-delivered motivational interventions to reduce drinking. Such efforts will require using advanced speech recognition capabilities and machine-learning approaches to train a program to mirror the decisions made by human controllers in the voice-based computer-delivered intervention used in this study. In addition, future studies should examine enhancements that can increase the perceived warmth and empathy of voice-based computer-delivered intervention, possibly through greater personalization, improvements in the speech generation software, and embodying the computer-delivered intervention in a physical form. PMID:28659259

  15. The Relationship between Personality Type and Acceptable Noise Levels: A Pilot Study.

    PubMed

    Franklin, Cliff; Johnson, Laura V; White, Letitia; Franklin, Clay; Smith-Olinde, Laura

    2013-01-01

    Objectives. This study examined the relationship between acceptable noise level (ANL) and personality. ANL is the difference between a person's most comfortable level for speech and the loudest level of background noise they are willing to accept while listening to speech. Design. Forty young adults with normal hearing participated. ANLs were measured and two personality tests (Big Five Inventory, Myers-Briggs Type Indicator) were administered. Results. The analysis revealed a correlation between ANL and the openness and conscientious personality dimensions from the Big Five Inventory; no correlation emerged between ANL and the Myers-Briggs personality types. Conclusions. Lower ANLs are correlated with full-time hearing aid use and the openness personality dimension; higher ANLs are correlated with part-time or hearing aid nonuse and the conscientious personality dimension. Current data suggest that those more open to new experiences may accept more noise and possibly be good hearing aid candidates, while those more conscientious may accept less noise and reject hearing aids, based on their unwillingness to accept background noise. Knowing something about a person's personality type may help audiologists determine if their patients will likely be good candidates for hearing aids.

  16. The Relationship between Personality Type and Acceptable Noise Levels: A Pilot Study

    PubMed Central

    Franklin, Cliff; Johnson, Laura V.; Franklin, Clay

    2013-01-01

    Objectives. This study examined the relationship between acceptable noise level (ANL) and personality. ANL is the difference between a person's most comfortable level for speech and the loudest level of background noise they are willing to accept while listening to speech. Design. Forty young adults with normal hearing participated. ANLs were measured and two personality tests (Big Five Inventory, Myers-Briggs Type Indicator) were administered. Results. The analysis revealed a correlation between ANL and the openness and conscientious personality dimensions from the Big Five Inventory; no correlation emerged between ANL and the Myers-Briggs personality types. Conclusions. Lower ANLs are correlated with full-time hearing aid use and the openness personality dimension; higher ANLs are correlated with part-time or hearing aid nonuse and the conscientious personality dimension. Current data suggest that those more open to new experiences may accept more noise and possibly be good hearing aid candidates, while those more conscientious may accept less noise and reject hearing aids, based on their unwillingness to accept background noise. Knowing something about a person's personality type may help audiologists determine if their patients will likely be good candidates for hearing aids. PMID:24349796

  17. Speech-on-speech masking with variable access to the linguistic content of the masker speech for native and nonnative english speakers.

    PubMed

    Calandruccio, Lauren; Bradlow, Ann R; Dhar, Sumitrajit

    2014-04-01

    Masking release for an English sentence-recognition task in the presence of foreign-accented English speech compared with native-accented English speech was reported in Calandruccio et al (2010a). The masking release appeared to increase as the masker intelligibility decreased. However, it could not be ruled out that spectral differences between the speech maskers were influencing the significant differences observed. The purpose of the current experiment was to minimize spectral differences between speech maskers to determine how various amounts of linguistic information within competing speech Affiliationect masking release. A mixed-model design with within-subject (four two-talker speech maskers) and between-subject (listener group) factors was conducted. Speech maskers included native-accented English speech and high-intelligibility, moderate-intelligibility, and low-intelligibility Mandarin-accented English. Normalizing the long-term average speech spectra of the maskers to each other minimized spectral differences between the masker conditions. Three listener groups were tested, including monolingual English speakers with normal hearing, nonnative English speakers with normal hearing, and monolingual English speakers with hearing loss. The nonnative English speakers were from various native language backgrounds, not including Mandarin (or any other Chinese dialect). Listeners with hearing loss had symmetric mild sloping to moderate sensorineural hearing loss. Listeners were asked to repeat back sentences that were presented in the presence of four different two-talker speech maskers. Responses were scored based on the key words within the sentences (100 key words per masker condition). A mixed-model regression analysis was used to analyze the difference in performance scores between the masker conditions and listener groups. Monolingual English speakers with normal hearing benefited when the competing speech signal was foreign accented compared with native accented, allowing for improved speech recognition. Various levels of intelligibility across the foreign-accented speech maskers did not influence results. Neither the nonnative English-speaking listeners with normal hearing nor the monolingual English speakers with hearing loss benefited from masking release when the masker was changed from native-accented to foreign-accented English. Slight modifications between the target and the masker speech allowed monolingual English speakers with normal hearing to improve their recognition of native-accented English, even when the competing speech was highly intelligible. Further research is needed to determine which modifications within the competing speech signal caused the Mandarin-accented English to be less effective with respect to masking. Determining the influences within the competing speech that make it less effective as a masker or determining why monolingual normal-hearing listeners can take advantage of these differences could help improve speech recognition for those with hearing loss in the future. American Academy of Audiology.

  18. Auditory Brainstem Implantation in Chinese Patients With Neurofibromatosis Type II: The Hong Kong Experience.

    PubMed

    Thong, Jiun Fong; Sung, John K K; Wong, Terence K C; Tong, Michael C F

    2016-08-01

    To describe our experience and outcomes of auditory brainstem implantation (ABI) in Chinese patients with Neurofibromatosis Type II (NF2). Retrospective case review. Tertiary referral center. Patients with NF2 who received ABIs. Between 1997 and 2014, eight patients with NF2 received 9 ABIs after translabyrinthine removal of their vestibular schwannomas. One patient did not have auditory response using the ABI after activation. Environmental sounds could be differentiated by six (75%) patients after 6 months of ABI use (mean score 46% [range 28-60%]), and by five (63%) patients after 1 year (mean score 57% [range 36-76%]) and 2 years of ABI use (mean score 48% [range 24-76%]). Closed-set word identification was possible in four (50%) patients after 6 months (mean score 39% [range 12-72%]), 1 year (mean score 68% [range 48-92%]), and 2 years of ABI use (mean score 62% [range 28-100%]). No patient demonstrated open-set sentence recognition in quiet in the ABI-only condition. However, the use of ABI together with lip-reading conferred an improvement over lip-reading alone in open-set sentence recognition scores in two (25%) patients after 6 months of ABI use (mean improvement 46%), and five (63%) patients after 1 year (mean improvement 25%) and 2 years of ABI use (mean improvement 28%). At 2 years postoperatively, three (38%) patients remained ABI users. This is the only published study to date examining ABI outcomes in Cantonese-speaking Chinese NF2 patients and the data seems to show poorer outcomes compared with English-speaking and other nontonal language-speaking NF2 patients. Environmental sound awareness and lip-reading enhancement are the main benefits observed in our patients. More work is needed to improve auditory implant speech-processing strategies for tonal languages and these advancements may yield better speech perception outcomes in the future.

  19. Designing augmentative and alternative communication applications: the results of focus groups with speech-language pathologists and parents of children with autism spectrum disorder.

    PubMed

    Boster, Jamie B; McCarthy, John W

    2018-05-01

    The purpose of this study was to gain insight from speech-language pathologists (SLPs) and parents of children with autism spectrum disorder (ASD) regarding appealing features of augmentative and alternative communication (AAC) applications. Two separate 1-hour focus groups were conducted with 8 SLPs and 5 parents of children with ASD to identify appealing design features of AAC Apps, their benefits and potential concerns. Participants were shown novel interface designs for communication mode, play mode and incentive systems. Participants responded to poll questions and provided benefits and drawbacks of the features as part of structured discussion. SLPs and parents identified a range of appealing features in communication mode (customization, animation and colour-coding) as well as in play mode (games and videos). SLPs preferred interfaces that supported motor planning and instruction while parents preferred those features such as character assistants that would appeal to their child. Overall SLPs and parents agreed on features for future AAC Apps. SLPs and parents have valuable input in regards to future AAC app design informed by their experiences with children with ASD. Both groups are key stakeholders in the design process and should be included in future design and research endeavors. Implications for Rehabilitation AAC applications for the iPad are often designed based on previous devices without consideration of new features. Ensuring the design of new interfaces are appealing and beneficial for children with ASD can potentially further support their communication. This study demonstrates how key stakeholders in AAC including speech language pathologists and parents can provide information to support the development of future AAC interface designs. Key stakeholders may be an untapped resource in the development of future AAC interfaces for children with ASD.

  20. Non-Deployables: An Increasing Challenge for the Army

    DTIC Science & Technology

    2010-04-01

    under criminal investigation for desertion, spying, espionage, aiding the enemy, treason, sedition , or subversive activities or who are under arrest...Armed Services Committee Testimony Opening Remarks (as delivered), February 25, 2010, http://www.army.mil/- speeches /2010/03/01/35133-feb-25- 2010

  1. Speech acts and performances of scientific citizenship: Examining how scientists talk about therapeutic cloning.

    PubMed

    Marks, Nicola J

    2014-07-01

    Scientists play an important role in framing public engagement with science. Their language can facilitate or impede particular interactions taking place with particular citizens: scientists' "speech acts" can "perform" different types of "scientific citizenship". This paper examines how scientists in Australia talked about therapeutic cloning during interviews and during the 2006 parliamentary debates on stem cell research. Some avoided complex labels, thereby facilitating public examination of this field. Others drew on language that only opens a space for publics to become educated, not to participate in a more meaningful way. Importantly, public utterances made by scientists here contrast with common international utterances: they did not focus on the therapeutic but the research promises of therapeutic cloning. Social scientists need to pay attention to the performative aspects of language in order to promote genuine citizen involvement in techno-science. Speech Act Theory is a useful analytical tool for this.

  2. Communication as group process media of aircrew performance

    NASA Technical Reports Server (NTRS)

    Kanki, B. G.; Foushee, H. C.

    1989-01-01

    This study of group process was motivated by a high-fidelity flight simulator project in which aircrew performance was found to be better when the crew had recently flown together. Considering recent operating experience as a group-level input factor, aspects of the communication process between crewmembers (Captain and First Officer), were explored as a possible mediator to performance. Communication patterns were defined by a speech act typology adapted for the flightdeck setting and distinguished crews that had previously flown together (FT) from those that had not flown together (NFT). A more open communication channel with respect to information exchange and validation and greater First Officer participation in task-related topics was shown by FT crews while NFT crews engaged in more non-task discourse, a speech mode less structured by roles and probably serving a more interpersonal function. Relationships between the speech categories themselves, representing linguistic, and role-related interdependencies provide guidelines for interpreting the primary findings.

  3. What makes an automated teller machine usable by blind users?

    PubMed

    Manzke, J M; Egan, D H; Felix, D; Krueger, H

    1998-07-01

    Fifteen blind and sighted subjects, who featured as a control group for acceptance, were asked for their requirements for automated teller machines (ATMs). Both groups also tested the usability of a partially operational ATM mock-up. This machine was based on an existing cash dispenser, providing natural speech output, different function menus and different key arrangements. Performance and subjective evaluation data of blind and sighted subjects were collected. All blind subjects were able to operate the ATM successfully. The implemented speech output was the main usability factor for them. The different interface designs did not significantly affect performance and subjective evaluation. Nevertheless, design recommendations can be derived from the requirement assessment. The sighted subjects were rather open for design modifications, especially the implementation of speech output. However, there was also a mismatch of the requirements of the two subject groups, mainly concerning the key arrangement.

  4. Unilateral Hearing Loss: Understanding Speech Recognition and Localization Variability - Implications for Cochlear Implant Candidacy

    PubMed Central

    Firszt, Jill B.; Reeder, Ruth M.; Holden, Laura K.

    2016-01-01

    Objectives At a minimum, unilateral hearing loss (UHL) impairs sound localization ability and understanding speech in noisy environments, particularly if the loss is severe to profound. Accompanying the numerous negative consequences of UHL is considerable unexplained individual variability in the magnitude of its effects. Identification of co-variables that affect outcome and contribute to variability in UHLs could augment counseling, treatment options, and rehabilitation. Cochlear implantation as a treatment for UHL is on the rise yet little is known about factors that could impact performance or whether there is a group at risk for poor cochlear implant outcomes when hearing is near-normal in one ear. The overall goal of our research is to investigate the range and source of variability in speech recognition in noise and localization among individuals with severe to profound UHL and thereby help determine factors relevant to decisions regarding cochlear implantation in this population. Design The present study evaluated adults with severe to profound UHL and adults with bilateral normal hearing. Measures included adaptive sentence understanding in diffuse restaurant noise, localization, roving-source speech recognition (words from 1 of 15 speakers in a 140° arc) and an adaptive speech-reception threshold psychoacoustic task with varied noise types and noise-source locations. There were three age-gender-matched groups: UHL (severe to profound hearing loss in one ear and normal hearing in the contralateral ear), normal hearing listening bilaterally, and normal hearing listening unilaterally. Results Although the normal-hearing-bilateral group scored significantly better and had less performance variability than UHLs on all measures, some UHL participants scored within the range of the normal-hearing-bilateral group on all measures. The normal-hearing participants listening unilaterally had better monosyllabic word understanding than UHLs for words presented on the blocked/deaf side but not the open/hearing side. In contrast, UHLs localized better than the normal hearing unilateral listeners for stimuli on the open/hearing side but not the blocked/deaf side. This suggests that UHLs had learned strategies for improved localization on the side of the intact ear. The UHL and unilateral normal hearing participant groups were not significantly different for speech-in-noise measures. UHL participants with childhood rather than recent hearing loss onset localized significantly better; however, these two groups did not differ for speech recognition in noise. Age at onset in UHL adults appears to affect localization ability differently than understanding speech in noise. Hearing thresholds were significantly correlated with speech recognition for UHL participants but not the other two groups. Conclusions Auditory abilities of UHLs varied widely and could be explained only in part by hearing threshold levels. Age at onset and length of hearing loss influenced performance on some, but not all measures. Results support the need for a revised and diverse set of clinical measures, including sound localization, understanding speech in varied environments and careful consideration of functional abilities as individuals with severe to profound UHL are being considered potential cochlear implant candidates. PMID:28067750

  5. Improved outcomes in auditory brainstem implantation with the use of near-field electrical compound action potentials.

    PubMed

    Mandalà, Marco; Colletti, Liliana; Colletti, Giacomo; Colletti, Vittorio

    2014-12-01

    To compare the outcomes (auditory threshold and open-set speech perception at 48-month follow-up) of a new near-field monitoring procedure, electrical compound action potential, on positioning the auditory brainstem implant electrode array on the surface of the cochlear nuclei versus the traditional far-field electrical auditory brainstem response. Retrospective study. Tertiary referral center. Among the 202 patients with auditory brainstem implants fitted and monitored with electrical auditory brainstem response during implant fitting, 9 also underwent electrical compound action potential recording. These subjects were matched retrospectively with a control group of 9 patients in whom only the electrical auditory brainstem response was recorded. Electrical compound action potentials were obtained using a cotton-wick recording electrode located near the surface of the cochlear nuclei and on several cranial nerves. Significantly lower potential thresholds were observed with the recording electrode located on the cochlear nuclei surface compared with the electrical auditory brainstem response (104.4 ± 32.5 vs 158.9 ± 24.2, P = .0030). Electrical brainstem response and compound action potentials identified effects on the neighboring cranial nerves on 3.2 ± 2.4 and 7.8 ± 3.2 electrodes, respectively (P = .0034). Open-set speech perception outcomes at 48-month follow-up had improved significantly in the near- versus far-field recording groups (78.9% versus 56.7%; P = .0051). Electrical compound action potentials during auditory brainstem implantation significantly improved the definition of the potential threshold and the number of auditory and extra-auditory waves generated. It led to the best coupling between the electrode array and cochlear nuclei, significantly improving the overall open-set speech perception. © American Academy of Otolaryngology—Head and Neck Surgery Foundation 2014.

  6. Long-term effectiveness of the SpeechEasy fluency-enhancement device.

    PubMed

    Gallop, Ronald F; Runyan, Charles M

    2012-12-01

    The SpeechEasy has been found to be an effective device for reduction of stuttering frequency for many people who stutter (PWS); published studies typically have compared stuttering reduction at initial fitting of the device to results achieved up to one year later. This study examines long-term effectiveness by examining whether effects of the SpeechEasy were maintained for longer periods, from 13 to 59 months. Results indicated no significant change for seven device users from post-fitting to the time of the study (t=-.074, p=.943); however, findings varied greatly on a case-by-case basis. Most notably, when stuttering frequency for eleven users and former users, prior to device fitting, was compared to current stuttering frequency while not wearing the device, the change over time was found to be statistically significant (t=2.851, p=.017), suggesting a carry-over effect of the device. There was no significant difference in stuttering frequency when users were wearing versus not wearing the device currently (t=1.949, p=0.92). Examinations of these results, as well as direction for future research, are described herein. The reader will be able to: (a) identify and briefly describe two types of altered auditory feedback which the SpeechEasy incorporates in order to help reduce stuttering; (b) describe the carry-over effect found in this study, suggest effectiveness associated with the device over a longer period of time than previously reported, as well as its implications, and (c) list factors that might be assessed in future research involving this device in order to more narrowly determine which prospective users are most likely to benefit from employing the SpeechEasy. Copyright © 2012 Elsevier Inc. All rights reserved.

  7. Acceptable noise level (ANL) with Danish and non-semantic speech materials in adult hearing-aid users.

    PubMed

    Olsen, Steen Østergaard; Lantz, Johannes; Nielsen, Lars Holme; Brännström, K Jonas

    2012-09-01

    The acceptable noise level (ANL) test is used for quantification of the amount of background noise subjects accept when listening to speech. This study investigates Danish hearing-aid users' ANL performance using Danish and non-semantic speech signals, the repeatability of ANL, and the association between ANL and outcome of the international outcome inventory for hearing aids (IOI-HA). ANL was measured in three conditions in both ears at two test sessions. Subjects completed the IOI-HA and the ANL questionnaire. Sixty-three Danish hearing-aid users; fifty-seven subjects were full time users and 6 were part time/non users of hearing aids according to the ANL questionnaire. ANLs were similar to results with American English speech material. The coefficient of repeatability (CR) was 6.5-8.8 dB. IOI-HA scores were not associated to ANL. Danish and non-semantic ANL versions yield results similar to the American English version. The magnitude of the CR indicates that ANL with Danish and non-semantic speech materials is not suitable for prediction of individual patterns of future hearing-aid use or evaluation of individual benefit from hearing-aid features. The ANL with Danish and non-semantic speech materials is not related to IOI-HA outcome.

  8. Some Behavioral and Neurobiological Constraints on Theories of Audiovisual Speech Integration: A Review and Suggestions for New Directions

    PubMed Central

    Altieri, Nicholas; Pisoni, David B.; Townsend, James T.

    2012-01-01

    Summerfield (1987) proposed several accounts of audiovisual speech perception, a field of research that has burgeoned in recent years. The proposed accounts included the integration of discrete phonetic features, vectors describing the values of independent acoustical and optical parameters, the filter function of the vocal tract, and articulatory dynamics of the vocal tract. The latter two accounts assume that the representations of audiovisual speech perception are based on abstract gestures, while the former two assume that the representations consist of symbolic or featural information obtained from visual and auditory modalities. Recent converging evidence from several different disciplines reveals that the general framework of Summerfield’s feature-based theories should be expanded. An updated framework building upon the feature-based theories is presented. We propose a processing model arguing that auditory and visual brain circuits provide facilitatory information when the inputs are correctly timed, and that auditory and visual speech representations do not necessarily undergo translation into a common code during information processing. Future research on multisensory processing in speech perception should investigate the connections between auditory and visual brain regions, and utilize dynamic modeling tools to further understand the timing and information processing mechanisms involved in audiovisual speech integration. PMID:21968081

  9. Normal Adult Aging and the Contextual Influences Affecting Speech and Meaningful Sound Perception

    PubMed Central

    Aydelott, Jennifer; Leech, Robert; Crinion, Jennifer

    2010-01-01

    It is widely accepted that hearing loss increases markedly with age, beginning in the fourth decade ISO 7029 (2000). Age-related hearing loss is typified by high-frequency threshold elevation and associated reductions in speech perception because speech sounds, especially consonants, become inaudible. Nevertheless, older adults often report additional and progressive difficulties in the perception and comprehension of speech, often highlighted in adverse listening conditions that exceed those reported by younger adults with a similar degree of high-frequency hearing loss (Dubno, Dirks, & Morgan) leading to communication difficulties and social isolation (Weinstein & Ventry). Some of the age-related decline in speech perception can be accounted for by peripheral sensory problems but cognitive aging can also be a contributing factor. In this article, we review findings from the psycholinguistic literature predominantly over the last four years and present a pilot study illustrating how normal age-related changes in cognition and the linguistic context can influence speech-processing difficulties in older adults. For significant progress in understanding and improving the auditory performance of aging listeners to be made, we discuss how future research will have to be much more specific not only about which interactions between auditory and cognitive abilities are critical but also how they are modulated in the brain. PMID:21307006

  10. Some behavioral and neurobiological constraints on theories of audiovisual speech integration: a review and suggestions for new directions.

    PubMed

    Altieri, Nicholas; Pisoni, David B; Townsend, James T

    2011-01-01

    Summerfield (1987) proposed several accounts of audiovisual speech perception, a field of research that has burgeoned in recent years. The proposed accounts included the integration of discrete phonetic features, vectors describing the values of independent acoustical and optical parameters, the filter function of the vocal tract, and articulatory dynamics of the vocal tract. The latter two accounts assume that the representations of audiovisual speech perception are based on abstract gestures, while the former two assume that the representations consist of symbolic or featural information obtained from visual and auditory modalities. Recent converging evidence from several different disciplines reveals that the general framework of Summerfield's feature-based theories should be expanded. An updated framework building upon the feature-based theories is presented. We propose a processing model arguing that auditory and visual brain circuits provide facilitatory information when the inputs are correctly timed, and that auditory and visual speech representations do not necessarily undergo translation into a common code during information processing. Future research on multisensory processing in speech perception should investigate the connections between auditory and visual brain regions, and utilize dynamic modeling tools to further understand the timing and information processing mechanisms involved in audiovisual speech integration.

  11. Automatic Speech Recognition Predicts Speech Intelligibility and Comprehension for Listeners With Simulated Age-Related Hearing Loss.

    PubMed

    Fontan, Lionel; Ferrané, Isabelle; Farinas, Jérôme; Pinquier, Julien; Tardieu, Julien; Magnen, Cynthia; Gaillard, Pascal; Aumont, Xavier; Füllgrabe, Christian

    2017-09-18

    The purpose of this article is to assess speech processing for listeners with simulated age-related hearing loss (ARHL) and to investigate whether the observed performance can be replicated using an automatic speech recognition (ASR) system. The long-term goal of this research is to develop a system that will assist audiologists/hearing-aid dispensers in the fine-tuning of hearing aids. Sixty young participants with normal hearing listened to speech materials mimicking the perceptual consequences of ARHL at different levels of severity. Two intelligibility tests (repetition of words and sentences) and 1 comprehension test (responding to oral commands by moving virtual objects) were administered. Several language models were developed and used by the ASR system in order to fit human performances. Strong significant positive correlations were observed between human and ASR scores, with coefficients up to .99. However, the spectral smearing used to simulate losses in frequency selectivity caused larger declines in ASR performance than in human performance. Both intelligibility and comprehension scores for listeners with simulated ARHL are highly correlated with the performances of an ASR-based system. In the future, it needs to be determined if the ASR system is similarly successful in predicting speech processing in noise and by older people with ARHL.

  12. Field-testing the new DECtalk PC system for medical applications

    NASA Technical Reports Server (NTRS)

    Grams, R. R.; Smillov, A.; Li, B.

    1992-01-01

    Synthesized human speech has now reached a new level of performance. With the introduction of DEC's new DECtalk PC, the small system developer will have a very powerful tool for creative design. It has been our privilege to be involved in the beta-testing of this new device and to add a medical dictionary which covers a wide range of medical terminology. With the inherent board level understanding of speech synthesis and the medical dictionary, it is now possible to provide full digital speech output for all medical files and terms. The application of these tools will cover a wide range of options for the future and allow a new dimension in dealing with the complex user interface experienced in medical practice.

  13. Transcranial direct current stimulation in post stroke aphasia and primary progressive aphasia: Current knowledge and future clinical applications.

    PubMed

    Sebastian, Rajani; Tsapkini, Kyrana; Tippett, Donna C

    2016-06-13

    The application of transcranial direct current stimulation (tDCS) in chronic post stroke aphasia is documented in a substantial literature, and there is some new evidence that tDCS can augment favorable language outcomes in primary progressive aphasia. Anodal tDCS is most often applied to the left hemisphere language areas to increase cortical excitability (increase the threshold of activation) and cathodal tDCS is most often applied to the right hemisphere homotopic areas to inhibit over activation in contralesional right homologues of language areas. Outcomes usually are based on neuropsychological and language test performance, following a medical model which emphasizes impairment of function, rather than a model which emphasizes functional communication. In this paper, we review current literature of tDCS as it is being used as a research tool, and discuss future implementation of tDCS as an adjuvant treatment to behavioral speech-language pathology intervention. We review literature describing non-invasive brain stimulation, the mechanism of tDCS, and studies of tDCS in aphasia and neurodegenerative disorders. We discuss future clinical applications. tDCS is a promising adjunct to traditional speech-language pathology intervention to address speech-language deficits after stroke and in the neurodegenerative disease, primary progressive aphasia. Limited data are available regarding how performance on these types of specific tasks translates to functional communication outcomes.

  14. Transcranial Direct Current Stimulation in Post Stroke Aphasia and Primary Progressive Aphasia: Current Knowledge and Future Clinical Applications

    PubMed Central

    Sebastian, Rajani; Tsapkini, Kyrana; Tippett, Donna C.

    2016-01-01

    BACKGROUND The application of transcranial direct current stimulation (tDCS) in chronic post stroke aphasia is documented in a substantial literature, and there is some new evidence that tDCS can augment favorable language outcomes in primary progressive aphasia. Anodal tDCS is most often applied to the left hemisphere language areas to increase cortical excitability (increase the threshold of activation) and cathodal tDCS is most often applied to the right hemisphere homotopic areas to inhibit over activation in contralesional right homologues of language areas. Outcomes usually are based on neuropsychological and language test performance, following a medical model which emphasizes impairment of function, rather than a model which emphasizes functional communication. OBJECTIVE In this paper, we review current literature of tDCS as it is being used as a research tool, and discuss future implementation of tDCS as an adjuvant treatment to behavioral speech-language pathology intervention. METHODS We review literature describing non-invasive brain stimulation, the mechanism of tDCS, and studies of tDCS in aphasia and neurodegenerative disorders. We discuss future clinical applications. RESULTS/CONCLUSIONS tDCS is a promising adjunct to traditional speech-language pathology intervention to address speech-language deficits after stroke and in the neurodegenerative disease, primary progressive aphasia. Limited data are available regarding how performance on these types of specific tasks translates to functional communication outcomes. PMID:27314871

  15. Text to Speech (TTS) Capabilities for the Common Driver Trainer (CDT)

    DTIC Science & Technology

    2010-10-01

    harnessing in’leigle jalClpeno jocelyn linu ~ los angeles lottery margarine mathematlze mathematized mathematized meme memes memol...including Julie, Kate, and Paul . Based upon the names of the voices, it may be that the VoiceText capability is the technology being used currently on...DFTTSExportToFileEx(O, " Paul ", 1, 1033, "Testing the Digital Future Text-to-Speech SDK.", -1, -1, -1, -1, -1, DFTTS_ TEXT_ TYPE_ XML, "test.wav", 0, "", -1

  16. The Miserable Truth.

    ERIC Educational Resources Information Center

    Shaughnessy, Mina

    1998-01-01

    Presents Mina Shaughnessy's 1976 speech, given at a time when great financial crisis threatened to result in the wholesale disestablishment of basic writing programs. Notes grave perils, and outlines some of the truths uncovered through the experience of Open Admissions in higher education, truths that are indestructible despite retrenchments and…

  17. Self-perception versus professional assessment of functional outcome after ablative surgery in patients with oral cancer.

    PubMed

    Beck-Broichsitter, Benedicta E; Huck, Jörn; Küchler, Thomas; Hauke, Daniela; Hedderich, Jürgen; Wiltfang, Jörg; Becker, Stephan T

    2017-02-01

    The extent of functional impairment after ablative surgery in the orofacial region may be directly reflected in a reduction in Quality of Life. This study intended to compare the patients' perception with an objective functional evaluation of the orofacial system in order to bilaterally distinguish direct influence factors. A total of 45 patients were included in this study and were asked to complete the European Organisation for Research and Treatment of Cancer Core Questionnaire (EORTC QLQ-30) and the Head and Neck 35 Module (H&N 35). Afterward one independent speech therapist evaluated the patients applying the Frenchay Dysarthria Assessment regarding four main categories: mouth opening, tongue motility, swallowing and intelligibility. Comparisons between groups were performed using Whitney-Mann U-Wilcoxon test and calculating Spearman's rho. Overall the professional assessments by the speech therapists revealed significantly higher scores regarding intelligibility, swallowing and mouth opening when compared to the patients' self-perception. Smaller tumor sizes, no bone resection and local reconstruction techniques led to significantly better functional outcomes, when assessed by speech therapists. Swallowing was perceived significantly better by patients in cases of local reconstruction. From the professionals' point of view differences were perceived in more items compared to the patients' self-assessments, who widely experienced a more severe functional impairment. Physicians should take this into account when discussing adverse therapy effects with the patients.

  18. Computer-assisted CI fitting: Is the learning capacity of the intelligent agent FOX beneficial for speech understanding?

    PubMed

    Meeuws, Matthias; Pascoal, David; Bermejo, Iñigo; Artaso, Miguel; De Ceulaer, Geert; Govaerts, Paul J

    2017-07-01

    The software application FOX ('Fitting to Outcome eXpert') is an intelligent agent to assist in the programing of cochlear implant (CI) processors. The current version utilizes a mixture of deterministic and probabilistic logic which is able to improve over time through a learning effect. This study aimed at assessing whether this learning capacity yields measurable improvements in speech understanding. A retrospective study was performed on 25 consecutive CI recipients with a median CI use experience of 10 years who came for their annual CI follow-up fitting session. All subjects were assessed by means of speech audiometry with open set monosyllables at 40, 55, 70, and 85 dB SPL in quiet with their home MAP. Other psychoacoustic tests were executed depending on the audiologist's clinical judgment. The home MAP and the corresponding test results were entered into FOX. If FOX suggested to make MAP changes, they were implemented and another speech audiometry was performed with the new MAP. FOX suggested MAP changes in 21 subjects (84%). The within-subject comparison showed a significant median improvement of 10, 3, 1, and 7% at 40, 55, 70, and 85 dB SPL, respectively. All but two subjects showed an instantaneous improvement in their mean speech audiometric score. Persons with long-term CI use, who received a FOX-assisted CI fitting at least 6 months ago, display improved speech understanding after MAP modifications, as recommended by the current version of FOX. This can be explained only by intrinsic improvements in FOX's algorithms, as they have resulted from learning. This learning is an inherent feature of artificial intelligence and it may yield measurable benefit in speech understanding even in long-term CI recipients.

  19. The contribution of visual information to the perception of speech in noise with and without informative temporal fine structure

    PubMed Central

    Stacey, Paula C.; Kitterick, Pádraig T.; Morris, Saffron D.; Sumner, Christian J.

    2017-01-01

    Understanding what is said in demanding listening situations is assisted greatly by looking at the face of a talker. Previous studies have observed that normal-hearing listeners can benefit from this visual information when a talker's voice is presented in background noise. These benefits have also been observed in quiet listening conditions in cochlear-implant users, whose device does not convey the informative temporal fine structure cues in speech, and when normal-hearing individuals listen to speech processed to remove these informative temporal fine structure cues. The current study (1) characterised the benefits of visual information when listening in background noise; and (2) used sine-wave vocoding to compare the size of the visual benefit when speech is presented with or without informative temporal fine structure. The accuracy with which normal-hearing individuals reported words in spoken sentences was assessed across three experiments. The availability of visual information and informative temporal fine structure cues was varied within and across the experiments. The results showed that visual benefit was observed using open- and closed-set tests of speech perception. The size of the benefit increased when informative temporal fine structure cues were removed. This finding suggests that visual information may play an important role in the ability of cochlear-implant users to understand speech in many everyday situations. Models of audio-visual integration were able to account for the additional benefit of visual information when speech was degraded and suggested that auditory and visual information was being integrated in a similar way in all conditions. The modelling results were consistent with the notion that audio-visual benefit is derived from the optimal combination of auditory and visual sensory cues. PMID:27085797

  20. Visual speech discrimination and identification of natural and synthetic consonant stimuli

    PubMed Central

    Files, Benjamin T.; Tjan, Bosco S.; Jiang, Jintao; Bernstein, Lynne E.

    2015-01-01

    From phonetic features to connected discourse, every level of psycholinguistic structure including prosody can be perceived through viewing the talking face. Yet a longstanding notion in the literature is that visual speech perceptual categories comprise groups of phonemes (referred to as visemes), such as /p, b, m/ and /f, v/, whose internal structure is not informative to the visual speech perceiver. This conclusion has not to our knowledge been evaluated using a psychophysical discrimination paradigm. We hypothesized that perceivers can discriminate the phonemes within typical viseme groups, and that discrimination measured with d-prime (d’) and response latency is related to visual stimulus dissimilarities between consonant segments. In Experiment 1, participants performed speeded discrimination for pairs of consonant-vowel spoken nonsense syllables that were predicted to be same, near, or far in their perceptual distances, and that were presented as natural or synthesized video. Near pairs were within-viseme consonants. Natural within-viseme stimulus pairs were discriminated significantly above chance (except for /k/-/h/). Sensitivity (d’) increased and response times decreased with distance. Discrimination and identification were superior with natural stimuli, which comprised more phonetic information. We suggest that the notion of the viseme as a unitary perceptual category is incorrect. Experiment 2 probed the perceptual basis for visual speech discrimination by inverting the stimuli. Overall reductions in d’ with inverted stimuli but a persistent pattern of larger d’ for far than for near stimulus pairs are interpreted as evidence that visual speech is represented by both its motion and configural attributes. The methods and results of this investigation open up avenues for understanding the neural and perceptual bases for visual and audiovisual speech perception and for development of practical applications such as visual lipreading/speechreading speech synthesis. PMID:26217249

  1. A new comparator account of auditory verbal hallucinations: how motor prediction can plausibly contribute to the sense of agency for inner speech

    PubMed Central

    Swiney, Lauren; Sousa, Paulo

    2014-01-01

    The comparator account holds that processes of motor prediction contribute to the sense of agency by attenuating incoming sensory information and that disruptions to this process contribute to misattributions of agency in schizophrenia. Over the last 25 years this simple and powerful model has gained widespread support not only as it relates to bodily actions but also as an account of misattributions of agency for inner speech, potentially explaining the etiology of auditory verbal hallucination (AVH). In this paper we provide a detailed analysis of the traditional comparator account for inner speech, pointing out serious problems with the specification of inner speech on which it is based and highlighting inconsistencies in the interpretation of the electrophysiological evidence commonly cited in its favor. In light of these analyses we propose a new comparator account of misattributed inner speech. The new account follows leading models of motor imagery in proposing that inner speech is not attenuated by motor prediction, but rather derived directly from it. We describe how failures of motor prediction would therefore directly affect the phenomenology of inner speech and trigger a mismatch in the comparison between motor prediction and motor intention, contributing to abnormal feelings of agency. We argue that the new account fits with the emerging phenomenological evidence that AVHs are both distinct from ordinary inner speech and heterogeneous. Finally, we explore the possibility that the new comparator account may extend to explain disruptions across a range of imagistic modalities, and outline avenues for future research. PMID:25221502

  2. A new comparator account of auditory verbal hallucinations: how motor prediction can plausibly contribute to the sense of agency for inner speech.

    PubMed

    Swiney, Lauren; Sousa, Paulo

    2014-01-01

    The comparator account holds that processes of motor prediction contribute to the sense of agency by attenuating incoming sensory information and that disruptions to this process contribute to misattributions of agency in schizophrenia. Over the last 25 years this simple and powerful model has gained widespread support not only as it relates to bodily actions but also as an account of misattributions of agency for inner speech, potentially explaining the etiology of auditory verbal hallucination (AVH). In this paper we provide a detailed analysis of the traditional comparator account for inner speech, pointing out serious problems with the specification of inner speech on which it is based and highlighting inconsistencies in the interpretation of the electrophysiological evidence commonly cited in its favor. In light of these analyses we propose a new comparator account of misattributed inner speech. The new account follows leading models of motor imagery in proposing that inner speech is not attenuated by motor prediction, but rather derived directly from it. We describe how failures of motor prediction would therefore directly affect the phenomenology of inner speech and trigger a mismatch in the comparison between motor prediction and motor intention, contributing to abnormal feelings of agency. We argue that the new account fits with the emerging phenomenological evidence that AVHs are both distinct from ordinary inner speech and heterogeneous. Finally, we explore the possibility that the new comparator account may extend to explain disruptions across a range of imagistic modalities, and outline avenues for future research.

  3. Recognizing emotional speech in Persian: a validated database of Persian emotional speech (Persian ESD).

    PubMed

    Keshtiari, Niloofar; Kuhlmann, Michael; Eslami, Moharram; Klann-Delius, Gisela

    2015-03-01

    Research on emotional speech often requires valid stimuli for assessing perceived emotion through prosody and lexical content. To date, no comprehensive emotional speech database for Persian is officially available. The present article reports the process of designing, compiling, and evaluating a comprehensive emotional speech database for colloquial Persian. The database contains a set of 90 validated novel Persian sentences classified in five basic emotional categories (anger, disgust, fear, happiness, and sadness), as well as a neutral category. These sentences were validated in two experiments by a group of 1,126 native Persian speakers. The sentences were articulated by two native Persian speakers (one male, one female) in three conditions: (1) congruent (emotional lexical content articulated in a congruent emotional voice), (2) incongruent (neutral sentences articulated in an emotional voice), and (3) baseline (all emotional and neutral sentences articulated in neutral voice). The speech materials comprise about 470 sentences. The validity of the database was evaluated by a group of 34 native speakers in a perception test. Utterances recognized better than five times chance performance (71.4 %) were regarded as valid portrayals of the target emotions. Acoustic analysis of the valid emotional utterances revealed differences in pitch, intensity, and duration, attributes that may help listeners to correctly classify the intended emotion. The database is designed to be used as a reliable material source (for both text and speech) in future cross-cultural or cross-linguistic studies of emotional speech, and it is available for academic research purposes free of charge. To access the database, please contact the first author.

  4. Toward a dual-learning systems model of speech category learning

    PubMed Central

    Chandrasekaran, Bharath; Koslov, Seth R.; Maddox, W. T.

    2014-01-01

    More than two decades of work in vision posits the existence of dual-learning systems of category learning. The reflective system uses working memory to develop and test rules for classifying in an explicit fashion, while the reflexive system operates by implicitly associating perception with actions that lead to reinforcement. Dual-learning systems models hypothesize that in learning natural categories, learners initially use the reflective system and, with practice, transfer control to the reflexive system. The role of reflective and reflexive systems in auditory category learning and more specifically in speech category learning has not been systematically examined. In this article, we describe a neurobiologically constrained dual-learning systems theoretical framework that is currently being developed in speech category learning and review recent applications of this framework. Using behavioral and computational modeling approaches, we provide evidence that speech category learning is predominantly mediated by the reflexive learning system. In one application, we explore the effects of normal aging on non-speech and speech category learning. Prominently, we find a large age-related deficit in speech learning. The computational modeling suggests that older adults are less likely to transition from simple, reflective, unidimensional rules to more complex, reflexive, multi-dimensional rules. In a second application, we summarize a recent study examining auditory category learning in individuals with elevated depressive symptoms. We find a deficit in reflective-optimal and an enhancement in reflexive-optimal auditory category learning. Interestingly, individuals with elevated depressive symptoms also show an advantage in learning speech categories. We end with a brief summary and description of a number of future directions. PMID:25132827

  5. Effects of Online Augmented Kinematic and Perceptual Feedback on Treatment of Speech Movements in Apraxia of Speech

    PubMed Central

    McNeil, M.R.; Katz, W.F.; Fossett, T.R.D.; Garst, D.M.; Szuminsky, N.J.; Carter, G.; Lim, K.Y.

    2010-01-01

    Apraxia of speech (AOS) is a motor speech disorder characterized by disturbed spatial and temporal parameters of movement. Research on motor learning suggests that augmented feedback may provide a beneficial effect for training movement. This study examined the effects of the presence and frequency of online augmented visual kinematic feedback (AVKF) and clinician-provided perceptual feedback on speech accuracy in 2 adults with acquired AOS. Within a single-subject multiple-baseline design, AVKF was provided using electromagnetic midsagittal articulography (EMA) in 2 feedback conditions (50 or 100%). Articulator placement was specified for speech motor targets (SMTs). Treated and baselined SMTs were in the initial or final position of single-syllable words, in varying consonant-vowel or vowel-consonant contexts. SMTs were selected based on each participant's pre-assessed erred productions. Productions were digitally recorded and online perceptual judgments of accuracy (including segment and intersegment distortions) were made. Inter- and intra-judge reliability for perceptual accuracy was high. Results measured by visual inspection and effect size revealed positive acquisition and generalization effects for both participants. Generalization occurred across vowel contexts and to untreated probes. Results of the frequency manipulation were confounded by presentation order. Maintenance of learned and generalized effects were demonstrated for 1 participant. These data provide support for the role of augmented feedback in treating speech movements that result in perceptually accurate speech production. Future investigations will explore the independent contributions of each feedback type (i.e. kinematic and perceptual) in producing efficient and effective training of SMTs in persons with AOS. PMID:20424468

  6. Caregiver and patient reported outcomes after repair of cleft lip and/or palate in the Philippines.

    PubMed

    Sharp, Helen M; Canady, John W; Ligot, Fernando A C; Hague, Rebecca A; Gutierrez, Johcille; Gutierrez, Johnny

    2008-03-01

    To establish the feasibility of conducting outcomes research among patients treated during a medical mission and to identify the salient outcomes for patients and caregivers in one region of the Philippines. Prospective structured interview conducted in or near participants' homes on the island of Cebu, Philippines. Individuals who had surgery at least 6 months earlier within the study region were eligible. Recipients of surgery were located from surgical records and by word of mouth. (1) Proportion of individuals located. (2) Primary outcomes of oral cleft repair identified for the sample. Of 99 people on a surgical list, 52 (53%) were located, eight were excluded (ineligible, unavailable, or inaccessible), and 44 agreed to participate in the study. Ten participants were identified via word of mouth. Seventy-five interviews were conducted (21 caregiver-patient pairs, one adolescent, and 32 caregivers of a child <7 years). Nearly all participants (99%) would encourage others to pursue surgery. Open-ended questions were coded to identify primary outcomes: improved speech (52%), improved eating (25%), social benefits (14%), and improved appearance (6%). Caregivers (50%) and patients (68%) reported that improved speech was the most important change after surgery. Patients and caregivers ascribe positive changes related to cleft surgery. Outcomes research requires cooperation with local professionals who can communicate effectively. These data serve to demonstrate feasibility and as a model for future studies of treatment outcomes in follow-up to international medical missions.

  7. Speech Recognition and Parent Ratings From Auditory Development Questionnaires in Children Who Are Hard of Hearing.

    PubMed

    McCreery, Ryan W; Walker, Elizabeth A; Spratford, Meredith; Oleson, Jacob; Bentler, Ruth; Holte, Lenore; Roush, Patricia

    2015-01-01

    Progress has been made in recent years in the provision of amplification and early intervention for children who are hard of hearing. However, children who use hearing aids (HAs) may have inconsistent access to their auditory environment due to limitations in speech audibility through their HAs or limited HA use. The effects of variability in children's auditory experience on parent-reported auditory skills questionnaires and on speech recognition in quiet and in noise were examined for a large group of children who were followed as part of the Outcomes of Children with Hearing Loss study. Parent ratings on auditory development questionnaires and children's speech recognition were assessed for 306 children who are hard of hearing. Children ranged in age from 12 months to 9 years. Three questionnaires involving parent ratings of auditory skill development and behavior were used, including the LittlEARS Auditory Questionnaire, Parents Evaluation of Oral/Aural Performance in Children rating scale, and an adaptation of the Speech, Spatial, and Qualities of Hearing scale. Speech recognition in quiet was assessed using the Open- and Closed-Set Test, Early Speech Perception test, Lexical Neighborhood Test, and Phonetically Balanced Kindergarten word lists. Speech recognition in noise was assessed using the Computer-Assisted Speech Perception Assessment. Children who are hard of hearing were compared with peers with normal hearing matched for age, maternal educational level, and nonverbal intelligence. The effects of aided audibility, HA use, and language ability on parent responses to auditory development questionnaires and on children's speech recognition were also examined. Children who are hard of hearing had poorer performance than peers with normal hearing on parent ratings of auditory skills and had poorer speech recognition. Significant individual variability among children who are hard of hearing was observed. Children with greater aided audibility through their HAs, more hours of HA use, and better language abilities generally had higher parent ratings of auditory skills and better speech-recognition abilities in quiet and in noise than peers with less audibility, more limited HA use, or poorer language abilities. In addition to the auditory and language factors that were predictive for speech recognition in quiet, phonological working memory was also a positive predictor for word recognition abilities in noise. Children who are hard of hearing continue to experience delays in auditory skill development and speech-recognition abilities compared with peers with normal hearing. However, significant improvements in these domains have occurred in comparison to similar data reported before the adoption of universal newborn hearing screening and early intervention programs for children who are hard of hearing. Increasing the audibility of speech has a direct positive effect on auditory skill development and speech-recognition abilities and also may enhance these skills by improving language abilities in children who are hard of hearing. Greater number of hours of HA use also had a significant positive impact on parent ratings of auditory skills and children's speech recognition.

  8. Cochlear Implantation in Older Adults

    PubMed Central

    Lin, Frank R.; Chien, Wade W.; Li, Lingsheng; Niparko, John K.; Francis, Howard W.

    2012-01-01

    Cochlear implants allow individuals with severe-to-profound hearing loss access to sound and spoken language. The number of older adults in the United States who are potential candidates for cochlear implantation is approximately 150,000 and will continue to increase with the aging of the population. Should cochlear implantation (CI) be routinely recommended for these older adults, and do these individuals benefit from CI? We reviewed our 12 year experience with cochlear implantation in adults ≥60 years (n = 445) at Johns Hopkins to investigate the impact of CI on speech understanding and to identify factors associated with speech performance. Complete data on speech outcomes at baseline and 1 year post-CI were available for 83 individuals. Our results demonstrate that cochlear implantation in adults ≥60 years consistently improved speech understanding scores with a mean increase of 60. 0% (S. D. 24. 1) on HINT sentences in quiet . The magnitude of the gain in speech scores was negatively associated with age at implantation such that for every increasing year of age at CI the gain in speech scores was 1. 3 percentage points less (95% CI: 0. 6 – 1. 9) after adjusting for age at hearing loss onset. Conversely, individuals with higher pre-CI speech scores (HINT scores between 40–60%) had significantly greater post-CI speech scores by a mean of 10. 0 percentage points (95% CI: 0. 4 – 19. 6) than those with lower pre-CI speech scores (HINT <40%) after adjusting for age at CI and age at hearing loss onset. These results suggest that older adult CI candidates who are younger at implantation and with higher preoperative speech scores obtain the highest speech understanding scores after cochlear implantation with possible implications for current Medicare policy. Finally, we provide an extended discussion of the epidemiology and impact of hearing loss in older adults. Future research of CI in older adults should expand beyond simple speech outcomes to take into account the broad cognitive, social, and physical functioning outcomes that are likely detrimentally impacted by hearing loss and may be mitigated by cochlear implantation. PMID:22932787

  9. Speech-on-speech masking with variable access to the linguistic content of the masker speech for native and non-native speakers of English

    PubMed Central

    Calandruccio, Lauren; Bradlow, Ann R.; Dhar, Sumitrajit

    2013-01-01

    Background Masking release for an English sentence-recognition task in the presence of foreign-accented English speech compared to native-accented English speech was reported in Calandruccio, Dhar and Bradlow (2010). The masking release appeared to increase as the masker intelligibility decreased. However, it could not be ruled out that spectral differences between the speech maskers were influencing the significant differences observed. Purpose The purpose of the current experiment was to minimize spectral differences between speech maskers to determine how various amounts of linguistic information within competing speech affect masking release. Research Design A mixed model design with within- (four two-talker speech maskers) and between-subject (listener group) factors was conducted. Speech maskers included native-accented English speech, and high-intelligibility, moderate-intelligibility and low-intelligibility Mandarin-accented English. Normalizing the long-term average speech spectra of the maskers to each other minimized spectral differences between the masker conditions. Study Sample Three listener groups were tested including monolingual English speakers with normal hearing, non-native speakers of English with normal hearing, and monolingual speakers of English with hearing loss. The non-native speakers of English were from various native-language backgrounds, not including Mandarin (or any other Chinese dialect). Listeners with hearing loss had symmetrical, mild sloping to moderate sensorineural hearing loss. Data Collection and Analysis Listeners were asked to repeat back sentences that were presented in the presence of four different two-talker speech maskers. Responses were scored based on the keywords within the sentences (100 keywords/masker condition). A mixed-model regression analysis was used to analyze the difference in performance scores between the masker conditions and the listener groups. Results Monolingual speakers of English with normal hearing benefited when the competing speech signal was foreign-accented compared to native-accented allowing for improved speech recognition. Various levels of intelligibility across the foreign-accented speech maskers did not influence results. Neither the non-native English listeners with normal hearing, nor the monolingual English speakers with hearing loss benefited from masking release when the masker was changed from native-accented to foreign-accented English. Conclusions Slight modifications between the target and the masker speech allowed monolingual speakers of English with normal hearing to improve their recognition of native-accented English even when the competing speech was highly intelligible. Further research is needed to determine which modifications within the competing speech signal caused the Mandarin-accented English to be less effective with respect to masking. Determining the influences within the competing speech that make it less effective as a masker, or determining why monolingual normal-hearing listeners can take advantage of these differences could help improve speech recognition for those with hearing loss in the future. PMID:25126683

  10. America on the Ice. Antarctic Policy Issues

    DTIC Science & Technology

    1990-01-01

    Prime Minister- Mahatir Mohamad-fired the open- ing volleys during a UN General Assembly speech in September of that year. He noted, "Henceforth all...problem of unin- habited lands." According to Mahatir , the Antarctic conti- nent clearly qualified for such consideration and, not withstanding the

  11. Unity Within Diversity.

    ERIC Educational Resources Information Center

    Kong, Shiu L., Ed.; Ray, Riten, Ed.

    This monograph contains the speeches and resolutions of representatives of forty-one ethnocultural groups residing in Canada. Included are: (1) opening messages by Pierre Trudeau, Joe Clark, Edward Broadbent, and Alex Chumak; (2) a discussion by Norman Cafik on federalism, multicultural groups, and Canadian unity; (3) remarks on Canadian unity by…

  12. Real-time interactive speech technology at Threshold Technology, Incorporated

    NASA Technical Reports Server (NTRS)

    Herscher, Marvin B.

    1977-01-01

    Basic real-time isolated-word recognition techniques are reviewed. Industrial applications of voice technology are described in chronological order of their development. Future research efforts are also discussed.

  13. Effect of age at cochlear implantation on auditory and speech development of children with auditory neuropathy spectrum disorder.

    PubMed

    Liu, Yuying; Dong, Ruijuan; Li, Yuling; Xu, Tianqiu; Li, Yongxin; Chen, Xueqing; Gong, Shusheng

    2014-12-01

    To evaluate the auditory and speech abilities in children with auditory neuropathy spectrum disorder (ANSD) after cochlear implantation (CI) and determine the role of age at implantation. Ten children participated in this retrospective case series study. All children had evidence of ANSD. All subjects had no cochlear nerve deficiency on magnetic resonance imaging and had used the cochlear implants for a period of 12-84 months. We divided our children into two groups: children who underwent implantation before 24 months of age and children who underwent implantation after 24 months of age. Their auditory and speech abilities were evaluated using the following: behavioral audiometry, the Categories of Auditory Performance (CAP), the Meaningful Auditory Integration Scale (MAIS), the Infant-Toddler Meaningful Auditory Integration Scale (IT-MAIS), the Standard-Chinese version of the Monosyllabic Lexical Neighborhood Test (LNT), the Multisyllabic Lexical Neighborhood Test (MLNT), the Speech Intelligibility Rating (SIR) and the Meaningful Use of Speech Scale (MUSS). All children showed progress in their auditory and language abilities. The 4-frequency average hearing level (HL) (500Hz, 1000Hz, 2000Hz and 4000Hz) of aided hearing thresholds ranged from 17.5 to 57.5dB HL. All children developed time-related auditory perception and speech skills. Scores of children with ANSD who received cochlear implants before 24 months tended to be better than those of children who received cochlear implants after 24 months. Seven children completed the Mandarin Lexical Neighborhood Test. Approximately half of the children showed improved open-set speech recognition. Cochlear implantation is helpful for children with ANSD and may be a good optional treatment for many ANSD children. In addition, children with ANSD fitted with cochlear implants before 24 months tended to acquire auditory and speech skills better than children fitted with cochlear implants after 24 months. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.

  14. Tracking development from early speech-language acquisition to reading skills at age 13.

    PubMed

    Bartl-Pokorny, Katrin D; Marschik, Peter B; Sachse, Steffi; Green, Vanessa A; Zhang, Dajie; Van Der Meer, Larah; Wolin, Thomas; Einspieler, Christa

    2013-06-01

    Previous studies have indicated a link between speech-language and literacy development. To add to this body of knowledge, we investigated whether lexical and grammatical skills from toddler to early school age are related to reading competence in adolescence. Twenty-three typically developing children were followed from age 1;6 to 13;6 (years;months). Parental checklists and standardized tests were used to assess the development of mental lexicon, grammatical and reading capacities of the children. Direct assessment of early speech-language functions positively correlated with later reading competence, whereas lexical skills reported by parents were not associated with this capacity. At (pre-) school age, larger vocabulary and better grammatical abilities predicted advanced reading abilities in adolescence. Our study contributes to the understanding of typical speech-language development and its relation to later reading outcome, extending the body of knowledge on these developmental domains for future early identification of children at risk for reading difficulties.

  15. Recent Advances in the Genetics of Vocal Learning

    PubMed Central

    Condro, Michael C.; White, Stephanie A.

    2015-01-01

    Language is a complex communicative behavior unique to humans, and its genetic basis is poorly understood. Genes associated with human speech and language disorders provide some insights, originating with the FOXP2 transcription factor, a mutation in which is the source of an inherited form of developmental verbal dyspraxia. Subsequently, targets of FOXP2 regulation have been associated with speech and language disorders, along with other genes. Here, we review these recent findings that implicate genetic factors in human speech. Due to the exclusivity of language to humans, no single animal model is sufficient to study the complete behavioral effects of these genes. Fortunately, some animals possess subcomponents of language. One such subcomponent is vocal learning, which though rare in the animal kingdom, is shared with songbirds. We therefore discuss how songbird studies have contributed to the current understanding of genetic factors that impact human speech, and support the continued use of this animal model for such studies in the future. PMID:26052371

  16. Effectiveness of the Directional Microphone in the Baha® Divino™

    PubMed Central

    Oeding, Kristi; Valente, Michael; Kerckhoff, Jessica

    2010-01-01

    Background Patients with unilateral sensorineural hearing loss (USNHL) experience great difficulty listening to speech in noisy environments. A directional microphone (DM) could potentially improve speech recognition in this difficult listening environment. It is well known that DMs in behind-the-ear (BTE) and custom hearing aids can provide a greater signal-to-noise ratio (SNR) in comparison to an omnidirectional microphone (OM) to improve speech recognition in noise for persons with hearing impairment. Studies examining the DM in bone anchored auditory osseointegrated implants (Baha), however, have been mixed, with little to no benefit reported for the DM compared to an OM. Purpose The primary purpose of this study was to determine if there are statistically significant differences in the mean reception threshold for sentences (RTS in dB) in noise between the OM and DM in the Baha® Divino™. The RTS of these two microphone modes was measured utilizing two loudspeaker arrays (speech from 0° and noise from 180° or a diffuse eight-loudspeaker array) and with the better ear open or closed with an earmold impression and noise attenuating earmuff. Subjective benefit was assessed using the Abbreviated Profile of Hearing Aid Benefit (APHAB) to compare unaided and aided (Divino OM and DM combined) problem scores. Research Design A repeated measures design was utilized, with each subject counterbalanced to each of the eight treatment levels for three independent variables: (1) microphone (OM and DM), (2) loudspeaker array (180° and diffuse), and (3) better ear (open and closed). Study Sample Sixteen subjects with USNHL currently utilizing the Baha were recruited from Washington University’s Center for Advanced Medicine and the surrounding area. Data Collection and Analysis Subjects were tested at the initial visit if they entered the study wearing the Divino or after at least four weeks of acclimatization to a loaner Divino. The RTS was determined utilizing Hearing in Noise Test (HINT) sentences in the R-Space™ system, and subjective benefit was determined utilizing the APHAB. A three-way repeated measures analysis of variance (ANOVA) and a paired samples t-test were utilized to analyze results of the HINT and APHAB, respectively. Results Results revealed statistically significant differences within microphone (p < 0.001; directional advantage of 3.2 dB), loudspeaker array (p = 0.046; 180° advantage of 1.1 dB), and better ear conditions (p < 0.001; open ear advantage of 4.9 dB). Results from the APHAB revealed statistically and clinically significant benefit for the Divino relative to unaided on the subscales of Ease of Communication (EC) (p = 0.037), Background Noise (BN) (p < 0.001), and Reverberation (RV) (p = 0.005). Conclusions The Divino’s DM provides a statistically significant improvement in speech recognition in noise compared to the OM for subjects with USNHL. Therefore, it is recommended that audiologists consider selecting a Baha with a DM to provide improved speech recognition performance in noisy listening environments. PMID:21034701

  17. Effectiveness of the directional microphone in the Baha® Divino™.

    PubMed

    Oeding, Kristi; Valente, Michael; Kerckhoff, Jessica

    2010-09-01

    Patients with unilateral sensorineural hearing loss (USNHL) experience great difficulty listening to speech in noisy environments. A directional microphone (DM) could potentially improve speech recognition in this difficult listening environment. It is well known that DMs in behind-the-ear (BTE) and custom hearing aids can provide a greater signal-to-noise ratio (SNR) in comparison to an omnidirectional microphone (OM) to improve speech recognition in noise for persons with hearing impairment. Studies examining the DM in bone anchored auditory osseointegrated implants (Baha), however, have been mixed, with little to no benefit reported for the DM compared to an OM. The primary purpose of this study was to determine if there are statistically significant differences in the mean reception threshold for sentences (RTS in dB) in noise between the OM and DM in the Baha® Divino™. The RTS of these two microphone modes was measured utilizing two loudspeaker arrays (speech from 0° and noise from 180° or a diffuse eight-loudspeaker array) and with the better ear open or closed with an earmold impression and noise attenuating earmuff. Subjective benefit was assessed using the Abbreviated Profile of Hearing Aid Benefit (APHAB) to compare unaided and aided (Divino OM and DM combined) problem scores. A repeated measures design was utilized, with each subject counterbalanced to each of the eight treatment levels for three independent variables: (1) microphone (OM and DM), (2) loudspeaker array (180° and diffuse), and (3) better ear (open and closed). Sixteen subjects with USNHL currently utilizing the Baha were recruited from Washington University's Center for Advanced Medicine and the surrounding area. Subjects were tested at the initial visit if they entered the study wearing the Divino or after at least four weeks of acclimatization to a loaner Divino. The RTS was determined utilizing Hearing in Noise Test (HINT) sentences in the R-Space™ system, and subjective benefit was determined utilizing the APHAB. A three-way repeated measures analysis of variance (ANOVA) and a paired samples t-test were utilized to analyze results of the HINT and APHAB, respectively. Results revealed statistically significant differences within microphone (p < 0.001; directional advantage of 3.2 dB), loudspeaker array (p = 0.046; 180° advantage of 1.1 dB), and better ear conditions (p < 0.001; open ear advantage of 4.9 dB). Results from the APHAB revealed statistically and clinically significant benefit for the Divino relative to unaided on the subscales of Ease of Communication (EC) (p = 0.037), Background Noise (BN) (p < 0.001), and Reverberation (RV) (p = 0.005). The Divino's DM provides a statistically significant improvement in speech recognition in noise compared to the OM for subjects with USNHL. Therefore, it is recommended that audiologists consider selecting a Baha with a DM to provide improved speech recognition performance in noisy listening environments. American Academy of Audiology.

  18. The role of voice input for human-machine communication.

    PubMed Central

    Cohen, P R; Oviatt, S L

    1995-01-01

    Optimism is growing that the near future will witness rapid growth in human-computer interaction using voice. System prototypes have recently been built that demonstrate speaker-independent real-time speech recognition, and understanding of naturally spoken utterances with vocabularies of 1000 to 2000 words, and larger. Already, computer manufacturers are building speech recognition subsystems into their new product lines. However, before this technology can be broadly useful, a substantial knowledge base is needed about human spoken language and performance during computer-based spoken interaction. This paper reviews application areas in which spoken interaction can play a significant role, assesses potential benefits of spoken interaction with machines, and compares voice with other modalities of human-computer interaction. It also discusses information that will be needed to build a firm empirical foundation for the design of future spoken and multimodal interfaces. Finally, it argues for a more systematic and scientific approach to investigating spoken input and performance with future language technology. PMID:7479803

  19. Speech Impairment in Down Syndrome: A Review

    PubMed Central

    Kent, Ray D.; Vorperian, Houri K.

    2012-01-01

    Purpose This review summarizes research on disorders of speech production in Down Syndrome (DS) for the purposes of informing clinical services and guiding future research. Method Review of the literature was based on searches using Medline, Google Scholar, Psychinfo, and HighWire Press, as well as consideration of reference lists in retrieved documents (including online sources). Search terms emphasized functions related to voice, articulation, phonology, prosody, fluency and intelligibility. Conclusions The following conclusions pertain to four major areas of review: (a) Voice. Although a number of studies have been reported on vocal abnormalities in DS, major questions remain about the nature and frequency of the phonatory disorder. Results of perceptual and acoustic studies have been mixed, making it difficult to draw firm conclusions or even to identify sensitive measures for future study. (b) Speech sounds. Articulatory and phonological studies show that speech patterns in DS are a combination of delayed development and errors not seen in typical development. Delayed (i.e., developmental) and disordered (i.e., nondevelopmental) patterns are evident by the age of about 3 years, although DS-related abnormalities possibly appear earlier, even in infant babbling. (c) Fluency and prosody. Stuttering and/or cluttering occur in DS at rates of 10 to 45%, compared to about 1% in the general population. Research also points to significant disturbances in prosody. (d) Intelligibility. Studies consistently show marked limitations in this area but it is only recently that research goes beyond simple rating scales. PMID:23275397

  20. Natural speech algorithm applied to baseline interview data can predict which patients will respond to psilocybin for treatment-resistant depression.

    PubMed

    Carrillo, Facundo; Sigman, Mariano; Fernández Slezak, Diego; Ashton, Philip; Fitzgerald, Lily; Stroud, Jack; Nutt, David J; Carhart-Harris, Robin L

    2018-04-01

    Natural speech analytics has seen some improvements over recent years, and this has opened a window for objective and quantitative diagnosis in psychiatry. Here, we used a machine learning algorithm applied to natural speech to ask whether language properties measured before psilocybin for treatment-resistant can predict for which patients it will be effective and for which it will not. A baseline autobiographical memory interview was conducted and transcribed. Patients with treatment-resistant depression received 2 doses of psilocybin, 10 mg and 25 mg, 7 days apart. Psychological support was provided before, during and after all dosing sessions. Quantitative speech measures were applied to the interview data from 17 patients and 18 untreated age-matched healthy control subjects. A machine learning algorithm was used to classify between controls and patients and predict treatment response. Speech analytics and machine learning successfully differentiated depressed patients from healthy controls and identified treatment responders from non-responders with a significant level of 85% of accuracy (75% precision). Automatic natural language analysis was used to predict effective response to treatment with psilocybin, suggesting that these tools offer a highly cost-effective facility for screening individuals for treatment suitability and sensitivity. The sample size was small and replication is required to strengthen inferences on these results. Copyright © 2018 Elsevier B.V. All rights reserved.

  1. Left-Dominant Temporal-Frontal Hypercoupling in Schizophrenia Patients With Hallucinations During Speech Perception

    PubMed Central

    Lavigne, Katie M.; Rapin, Lucile A.; Metzak, Paul D.; Whitman, Jennifer C.; Jung, Kwanghee; Dohen, Marion; Lœvenbruck, Hélène; Woodward, Todd S.

    2015-01-01

    Background: Task-based functional neuroimaging studies of schizophrenia have not yet replicated the increased coordinated hyperactivity in speech-related brain regions that is reported with symptom-capture and resting-state studies of hallucinations. This may be due to suboptimal selection of cognitive tasks. Methods: In the current study, we used a task that allowed experimental manipulation of control over verbal material and compared brain activity between 23 schizophrenia patients (10 hallucinators, 13 nonhallucinators), 22 psychiatric (bipolar), and 27 healthy controls. Two conditions were presented, one involving inner verbal thought (in which control over verbal material was required) and another involving speech perception (SP; in which control verbal material was not required). Results: A functional connectivity analysis resulted in a left-dominant temporal-frontal network that included speech-related auditory and motor regions and showed hypercoupling in past-week hallucinating schizophrenia patients (relative to nonhallucinating patients) during SP only. Conclusions: These findings replicate our previous work showing generalized speech-related functional network hypercoupling in schizophrenia during inner verbal thought and SP, but extend them by suggesting that hypercoupling is related to past-week hallucination severity scores during SP only, when control over verbal material is not required. This result opens the possibility that practicing control over inner verbal thought processes may decrease the likelihood or severity of hallucinations. PMID:24553150

  2. Value of pre-operative caloric test in predicting speech perception after cochlear implantation in adults with post-lingual hearing loss.

    PubMed

    Yang, Chan Joo; Lee, Jee Yeon; Ahn, Joong Ho; Lee, Kwang-Sun

    2016-09-01

    This study shows that, in cochlear implantation (CI) surgery, pre-operative caloric test results are not correlated with post-operative outcomes of dizziness or speech perception. To determine the role of pre-operative caloric tests in CI. The records of 95 patients who underwent unilateral CI were reviewed retrospectively. Patients were divided into four groups according to caloric response. Forty-six patients with normal caloric responses were classified as Group A, 19 patients who underwent CI in the ear with worse caloric responses as Group B, 18 patients with bilateral loss of caloric responses as Group C, and 12 patients who underwent CI in the ear with better caloric responses as Group D. Speech performance and post-operative dizziness were compared between the four groups. Speech perception was determined by evaluating consonant-vowel phoneme detection, closed-set word and open-set mono-syllabic and bi-syllabic word identification, and sentence comprehension test scores. The speech perception and aided pure-tone average (PTA) test results at 3 and 6 months and at 1, 2, and 3 years after implantation were not significantly different between Groups A, B, C, and D (p > 0.05). Eight patients (8.4%) reported post-operative dizziness, but there was no significant difference between the four groups (p = 0.627).

  3. A Construction System for CALL Materials from TV News with Captions

    NASA Astrophysics Data System (ADS)

    Kobayashi, Satoshi; Tanaka, Takashi; Mori, Kazumasa; Nakagawa, Seiichi

    Many language learning materials have been published. In language learning, although repetition training is obviously necessary, it is difficult to maintain the learner's interest/motivation using existing learning materials, because those materials are limited in their scope and contents. In addition, we doubt whether the speech sounds used in most materials are natural in various situations. Nowadays, some TV news programs (CNN, ABC, PBS, NHK, etc.) have closed/open captions corresponding to the announcer's speech. We have developed a system that makes Computer Assisted Language Learning (CALL) materials for both English learning by Japanese and Japanese learning by foreign students from such captioned newscasts. This system computes the synchronization between captions and speech by using HMMs and a forced alignment algorithm. Materials made by the system have following functions: full/partial text caption display, repetition listening, consulting an electronic dictionary, display of the user's/announcer's sound waveform and pitch contour, and automatic construction of a dictation test. Materials have following advantages: materials present polite and natural speech, various and timely topics. Furthermore, the materials have the following possibility: automatic creation of listening/understanding tests, and storage/retrieval of the many materials. In this paper, firstly, we present the organization of the system. Then, we describe results of questionnaires on trial use of the materials. As the result, we got enough accuracy on the synchronization between captions and speech. Speaking totally, we encouraged to research this system.

  4. Hypoglossal canal size and hominid speech

    PubMed Central

    DeGusta, David; Gilbert, W. Henry; Turner, Scott P.

    1999-01-01

    The mammalian hypoglossal canal transmits the nerve that supplies the motor innervation to the tongue. Hypoglossal canal size has previously been used to date the origin of human-like speech capabilities to at least 400,000 years ago and to assign modern human vocal abilities to Neandertals. These conclusions are based on the hypothesis that the size of the hypoglossal canal is indicative of speech capabilities. This hypothesis is falsified here by the finding of numerous nonhuman primate taxa that have hypoglossal canals in the modern human size range, both absolutely and relative to oral cavity volume. Specimens of Australopithecus afarensis, Australopithecus africanus, and Australopithecus boisei also have hypoglossal canals that, both absolutely and relative to oral cavity volume, are equal in size to those of modern humans. The basis for the hypothesis that hypoglossal canal size is indicative of speech was the assumption that hypoglossal canal size is correlated with hypoglossal nerve size, which in turn is related to tongue function. This assumption is probably incorrect, as we found no apparent correlation between the size of the hypoglossal nerve, or the number of axons it contains, and the size of the hypoglossal canal in a sample of cadavers. Our data demonstrate that the size of the hypoglossal canal does not reflect vocal capabilities or language usage. Thus the date of origin for human language and the speech capabilities of Neandertals remain open questions. PMID:9990105

  5. Speech intelligibility in hospitals.

    PubMed

    Ryherd, Erica E; Moeller, Michael; Hsu, Timothy

    2013-07-01

    Effective communication between staff members is key to patient safety in hospitals. A variety of patient care activities including admittance, evaluation, and treatment rely on oral communication. Surprisingly, published information on speech intelligibility in hospitals is extremely limited. In this study, speech intelligibility measurements and occupant evaluations were conducted in 20 units of five different U.S. hospitals. A variety of unit types and locations were studied. Results show that overall, no unit had "good" intelligibility based on the speech intelligibility index (SII > 0.75) and several locations found to have "poor" intelligibility (SII < 0.45). Further, occupied spaces were found to have 10%-15% lower SII than unoccupied spaces on average. Additionally, staff perception of communication problems at nurse stations was significantly correlated with SII ratings. In a targeted second phase, a unit treated with sound absorption had higher SII ratings for a larger percentage of time as compared to an identical untreated unit. Taken as a whole, the study provides an extensive baseline evaluation of speech intelligibility across a variety of hospitals and unit types, offers some evidence of the positive impact of absorption on intelligibility, and identifies areas for future research.

  6. [Development and equivalence evaluation of spondee lists of mandarin speech test materials].

    PubMed

    Zhang, Hua; Wang, Shuo; Wang, Liang; Chen, Jing; Chen, Ai-ting; Guo, Lian-sheng; Zhao, Xiao-yan; Ji, Chen

    2006-06-01

    To edit the spondee (disyllable) word lists as a part of mandarin speech test materials (MSTM). These will be basic speech materials for routine tests in clinics and laboratories. Two groups of professionals (audiologists, Chinese and Mandarin scientists, linguistician and statistician) were set up at first. The editing principles were established after 3 round table meetings. Ten spondee lists, each with 50 words, were edited and recorded into cassettes. All lists were phonemically balanced (3-dimensions: vowels, consonants and Chinese tones). Seventy-three normal hearing college students were tested. The speech was presented by earphone monaurally. Three statistic methods were used for equivalent analysis. Related analysis showed that all lists were much related, except List 5. Cluster analysis showed that all ten lists could be classified as two groups. But Kappa test showed that the lists' homogeneity were not well. Spondee lists are one of the most routine speech test materials. Their editing, recording and equivalent evaluation are affected by many factors. This also needs multi-discipline cooperation. All lists edited in present study need future modification in recording and testing in order to be used clinically and in research. The phonemic balance should be kept.

  7. Military and government applications of human-machine communication by voice.

    PubMed Central

    Weinstein, C J

    1995-01-01

    This paper describes a range of opportunities for military and government applications of human-machine communication by voice, based on visits and contacts with numerous user organizations in the United States. The applications include some that appear to be feasible by careful integration of current state-of-the-art technology and others that will require a varying mix of advances in speech technology and in integration of the technology into applications environments. Applications that are described include (1) speech recognition and synthesis for mobile command and control; (2) speech processing for a portable multifunction soldier's computer; (3) speech- and language-based technology for naval combat team tactical training; (4) speech technology for command and control on a carrier flight deck; (5) control of auxiliary systems, and alert and warning generation, in fighter aircraft and helicopters; and (6) voice check-in, report entry, and communication for law enforcement agents or special forces. A phased approach for transfer of the technology into applications is advocated, where integration of applications systems is pursued in parallel with advanced research to meet future needs. Images Fig. 1 Fig. 2 Fig. 3 Fig. 4 Fig. 5 Fig. 6 PMID:7479718

  8. Inequalities in the provision of paediatric speech and language therapy services across London boroughs.

    PubMed

    Pring, Tim

    2016-07-01

    The inverse-care law suggests that fewer healthcare resources are available in deprived areas where health needs are greatest. To examine the provision of paediatric speech and language services across London boroughs and to relate provision to the level of deprivation of the boroughs. Information on the employment of paediatric speech and language therapists was obtained from London boroughs by freedom-of-information requests. The relationship between the number of therapists and the index of multiple deprivation for the borough was examined. Twenty-nine of 32 boroughs responded. A positive relationship between provision and need was obtained, suggesting that the inverse-care law does not apply. However, large inequalities of provision were found particularly among the more socially deprived boroughs. In some instances boroughs had five times as many therapists per child as other boroughs. The data reveal that large differences in speech and language therapy provision exist across boroughs. The reasons for these inequalities are unclear, but the lack of comparative information across boroughs is likely to be unhelpful in planning equitable services. The use of freedom of information in assessing health inequalities is stressed and its future availability is desirable. © 2016 Royal College of Speech and Language Therapists.

  9. A Spoonful of (Added) Sugar Helps the Constitution Go Down: Curing the Compelled Speech Commercial Speech Doctrine with FDA’s Added Sugars Rule.

    PubMed

    Smith, Colleen

    On May 27, 2016, the Food and Drug Administration (FDA) announced that it was adopting a new rule that requires food manufacturers to list—on the already mandated Nutrition Facts label—how many grams of sugar have been added to a food product. Many opponents have criticized this “added sugars” rule on First Amendment grounds, arguing that the rule violates the commercial speech rights of food manufacturers. Whether the rule would survive constitutional scrutiny or not is an open question because the compelled commercial speech doctrine is anything but clear. Courts are split over whether Zauderer’s rational basis test, Central Hudson’s intermediate scrutiny, or some combination of the two should apply to a mandated disclosure like FDA’s added sugars rule. This Paper explains that the added sugars rule is unique in the history of mandated nutrition labeling in that the rule is motivated largely by public health concerns and backed by reports that assert that consumers should limit their intake of added sugars. In contrast, correcting and preventing consumer deception has been a major driving force behind the remainder of FDA’s mandated nutrition labeling. Because of this distinct rationale, the added sugars rule does not fit neatly into any currently existing compelled commercial speech test. This Paper uses the added sugars rule to highlight the deficiencies in the existing tests. Finally, this Paper proposes a new compelled commercial speech test that would adequately balance the interest of all of the effected parties: the government, the public, and food manufacturers.

  10. MMSE Estimator for Children’s Speech with Car and Weather Noise

    NASA Astrophysics Data System (ADS)

    Sayuthi, V.

    2018-04-01

    Previous research mentioned that most people need and use vehicles for various purposes, in this recent time and future, as a means of traveling. Many ways can be done in a vehicle, such as for enjoying entertainment, and doing work, so vehicles not just only as a means of traveling. In this study, we will examine the children’s speech from a girl in the vehicle that affected by noise disturbances from the sound source of car noise and the weather sound noise around it, in this case, the rainy weather noise. Vehicle sounds may be from car engine or car air conditioner. The minimum mean square error (MMSE) estimator is used as an attempt to obtain or detect the children’s clear speech by representing simulation research as random process signal that factored by the autocorrelation of both the child’s voice and the disturbance noise signal. This MMSE estimator can be considered as wiener filter as the clear sound are reconstructed again. We expected that the results of this study can help as the basis for development of entertainment or communication technology for passengers of vehicles in the future, particularly using MMSE estimators.

  11. Love and Politics: Sister Quinlan and the Future We Have Desired

    ERIC Educational Resources Information Center

    Ndebele, Njabulo S.

    2014-01-01

    Seamus Heaney (like Stimpson and Parker in this issue) speaks of: "A historic dialectic [which] exists between the beautiful and the bestial". In this speech, delivered on 13 December 2012, Njabulo Ndebele reflects on the stories South Africa tells itself about past atrocity, as a way of achieving "the future we have desired".…

  12. Researching Possible Futures to Guide Leaders towards More Effective Tertiary Education

    ERIC Educational Resources Information Center

    Davis, Niki; Higgins, Andrew

    2015-01-01

    This research aimed to inform institutional leaders by producing and disseminating a system wide view of what tertiary education might look like in Aotearoa New Zealand, five years into the future. The researchers were responding to a challenge in a speech at the DEANZ 2010 conference by a highly respected national leader (Dr. Peter Coolbear). The…

  13. Speech and Language: Translating the Genome.

    PubMed

    Deriziotis, Pelagia; Fisher, Simon E

    2017-09-01

    Investigation of the biological basis of human speech and language is being transformed by developments in molecular technologies, including high-throughput genotyping and next-generation sequencing of whole genomes. These advances are shedding new light on the genetic architecture underlying language-related disorders (speech apraxia, specific language impairment, developmental dyslexia) as well as that contributing to variation in relevant skills in the general population. We discuss how state-of-the-art methods are uncovering a range of genetic mechanisms, from rare mutations of large effect to common polymorphisms that increase risk in a subtle way, while converging on neurogenetic pathways that are shared between distinct disorders. We consider the future of the field, highlighting the unusual challenges and opportunities associated with studying genomics of language-related traits. Copyright © 2017 Elsevier Ltd. All rights reserved.

  14. Communicating with Virtual Humans.

    ERIC Educational Resources Information Center

    Thalmann, Nadia Magnenat

    The face is a small part of a human, but it plays an essential role in communication. An open hybrid system for facial animation is presented. It encapsulates a considerable amount of information regarding facial models, movements, expressions, emotions, and speech. The complex description of facial animation can be handled better by assigning…

  15. The Importance of Knowing Ourselves.

    ERIC Educational Resources Information Center

    Farquhar, Robin H.

    This speech focuses on the protection of institutional autonomy in higher education. The speaker opens by offering some of his own basic beliefs. The balance of the paper considers three key issues central to institutional autonomy. The first issue involves specific threats to autonomy: the socio-cultural mistrust of authority and institutions,…

  16. Beyond "It Was Great"? Not so Fast!

    ERIC Educational Resources Information Center

    Wong, E. David

    2015-01-01

    The Forum on Education Abroad's Ninth Annual Conference in 2013 was organized around the provocative theme, "Moving Beyond It Was Great." In the opening plenary speech, Lilli Engle warned the audience of study abroad researchers, leaders, administrators, and providers that study abroad programs were not as effective as they may want to…

  17. Utilizing Multi-Modal Literacies in Middle Grades Science

    ERIC Educational Resources Information Center

    Saurino, Dan; Ogletree, Tamra; Saurino, Penelope

    2010-01-01

    The nature of literacy is changing. Increased student use of computer-mediated, digital, and visual communication spans our understanding of adolescent multi-modal capabilities that reach beyond the traditional conventions of linear speech and written text in the science curriculum. Advancing technology opens doors to learning that involve…

  18. DefenseLink.mil - Special Report - Travels With Gates

    Science.gov Websites

    Hamad Al-Khalifa, during the International Institute for Stategic Studies' fifth annual Manama Dialogue Hamad Al-Khalifa, during the International Institute for Stategic Studies' fifth annual Manama Dialogue Secretary Robert M. Gates gives a speech during the opening of International Institute for Stategic Studies

  19. The Lice, Turkey, earthquake of September 6, 1975; a preliminary engineering investigation

    USGS Publications Warehouse

    Yanev, P. I.

    1976-01-01

    The Fifth European Conference on Earthquake Engineering was held on September 22 through 25 in Istanbul, Turkey. The opening speech by the Honorable H. E. Nurettin Ok, Minister of Reconstruction and Resettlement of Turkey, introduced the several hundred delegates to the realities of earthquake hazards in Turkey:

  20. Understanding speech in noise after correction of congenital unilateral aural atresia: effects of age in the emergence of binaural squelch but not in use of head-shadow.

    PubMed

    Gray, Lincoln; Kesser, Bradley; Cole, Erika

    2009-09-01

    Unilateral hearing loss causes difficulty hearing in noise (the "cocktail party effect") due to absence of redundancy, head-shadow, and binaural squelch. This study explores the emergence of the head-shadow and binaural squelch effects in children with unilateral congenital aural atresia undergoing surgery to correct their hearing deficit. Adding patients and data from a similar study previously published, we also evaluate covariates such as the age of the patient, surgical outcome, and complexity of the task that might predict the extent of binaural benefit--patients' ability to "use" their new ear--when understanding speech in noise. Patients with unilateral congenital aural atresia were tested for their ability to understand speech in noise before and again 1 month after surgery to repair their atresia. In a sound-attenuating booth participants faced a speaker that produced speech signals with noise 90 degrees to the side of the normal (non-atretic) ear and again to the side of the atretic ear. The Hearing in Noise Test (HINT for adults or HINT-C for children) was used to estimate the patients' speech reception thresholds. The speech-in-noise test (SPIN) or the Pediatric Speech Intelligibility (PSI) Test was used in the previous study. There was consistent improvement, averaging 5dB regardless of age, in the ability to take advantage of head-shadow in understanding speech with noise to the side of the non-atretic (normal) ear. There was, in contrast, a strong negative linear effect of age (r(2)=.78, selecting patients over 8 years) in the emergence of binaural squelch to understand speech with noise to the side of the atretic ear. In patients over 8 years, this trend replicated over different studies and different tests. Children less than 8 years, however, showed less improvement in the HINT-C than in the PSI after surgery with noise toward their atretic ear (effect size=3). No binaural result was correlated with degree of hearing improvement after surgery. All patients are able to take advantage of a favorable signal-to-noise ratio in their newly opened ear; that is with noise toward the side of the normal ear (but this physical, bilateral, head-shadow effect need not involve true central binaural processing). With noise toward the atretic ear, the emergence of binaural squelch replicates between two studies for all but the youngest patients. Approximately 2dB of binaural gain is lost for each decade that surgery is delayed, and zero (or poorer) binaural benefit is predicted after 38 years of age. Older adults do more poorly, possibly secondary to their long period of auditory deprivation. At the youngest ages, however, binaural results are different in open- and closed-set speech tests; the more complex hearing tasks may involve a greater cognitive load. Other cognitive abilities (late evoked potentials, grey matter in auditory cortex, and multitasking) show similar effects of age, peaking at the same late-teen/young-adult period. Longer follow-up is likely critical for the understanding of these data. Getting a new ear may be--like multitasking--challenging for the youngest and oldest subjects.

  1. Sequencing Complex Genomic Regions

    ScienceCinema

    Eichler, Evan

    2018-02-12

    Evan Eichler, Howard Hughes Medical Investigator at the University of Washington, gives the May 28, 2009 keynote speech at the "Sequencing, Finishing, Analysis in the Future" meeting in Santa Fe, NM. Part 1 of 2

  2. As You Get Older

    MedlinePlus

    ... to the Second Edition: Kathleen Kapp-Simon, PhD, Psychology David J. Reisberg, DDS, Prosthodontics James Sidman, MD, ... Speech ...........................................................................................................................8 Your Ears, Nose and Throat .......................................................................................10 Social Relationships ..........................................................................................................11 Looking Toward the Future ........................................................................................13 Stories ...

  3. New Ideas for Speech Recognition and Related Technologies

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Holzrichter, J F

    The ideas relating to the use of organ motion sensors for the purposes of speech recognition were first described by.the author in spring 1994. During the past year, a series of productive collaborations between the author, Tom McEwan and Larry Ng ensued and have lead to demonstrations, new sensor ideas, and algorithmic descriptions of a large number of speech recognition concepts. This document summarizes the basic concepts of recognizing speech once organ motions have been obtained. Micro power radars and their uses for the measurement of body organ motions, such as those of the heart and lungs, have been demonstratedmore » by Tom McEwan over the past two years. McEwan and I conducted a series of experiments, using these instruments, on vocal organ motions beginning in late spring, during which we observed motions of vocal folds (i.e., cords), tongue, jaw, and related organs that are very useful for speech recognition and other purposes. These will be reviewed in a separate paper. Since late summer 1994, Lawrence Ng and I have worked to make many of the initial recognition ideas more rigorous and to investigate the applications of these new ideas to new speech recognition algorithms, to speech coding, and to speech synthesis. I introduce some of those ideas in section IV of this document, and we describe them more completely in the document following this one, UCRL-UR-120311. For the design and operation of micro-power radars and their application to body organ motions, the reader may contact Tom McEwan directly. The capability for using EM sensors (i.e., radar units) to measure body organ motions and positions has been available for decades. Impediments to their use appear to have been size, excessive power, lack of resolution, and lack of understanding of the value of organ motion measurements, especially as applied to speech related technologies. However, with the invention of very low power, portable systems as demonstrated by McEwan at LLNL researchers have begun to think differently about practical applications of such radars. In particular, his demonstrations of heart and lung motions have opened up many new areas of application for human and animal measurements.« less

  4. Prevalence of swallowing and speech problems in daily life after chemoradiation for head and neck cancer based on cut-off scores of the patient-reported outcome measures SWAL-QOL and SHI.

    PubMed

    Rinkel, Rico N; Verdonck-de Leeuw, Irma M; Doornaert, Patricia; Buter, Jan; de Bree, Remco; Langendijk, Johannes A; Aaronson, Neil K; Leemans, C René

    2016-07-01

    The objective of this study is to assess swallowing and speech outcome after chemoradiation therapy for head and neck cancer, based on the patient-reported outcome measures Swallowing Quality of Life Questionnaire (SWAL-QOL) and Speech Handicap Index (SHI), both provided with cut-off scores. This is a cross-sectional study. Department of Otolaryngology/Head and Neck Surgery of a University Medical Center. Sixty patients, 6 months to 5 years after chemoradiation for head and neck squamous cell carcinoma. Swallowing Quality of Life Questionnaire (SWAL-QOL) and SHI, both validated in Dutch and provided with cut-off scores. Associations were tested between the outcome measures and independent variables (age, gender, tumor stage and site, and radiotherapy technique, time since treatment, comorbidity and food intake). Fifty-two patients returned the SWAL-QOL and 47 the SHI (response rate 87 and 78 %, respectively). Swallowing and speech problems were present in 79 and 55 %, respectively. Normal food intake was noticed in 45, 35 % had a soft diet and 20 % tube feeding. Patients with soft diet and tube feeding reported more swallowing problems compared to patients with normal oral intake. Tumor subsite was significantly associated with swallowing outcome (less problems in larynx/hypopharynx compared to oral/oropharynx). Radiation technique was significantly associated with psychosocial speech problems (less problems in patients treated with IMRT). Swallowing and (to a lesser extent) speech problems in daily life are frequently present after chemoradiation therapy for head and neck cancer. Future prospective studies will give more insight into the course of speech and swallowing problems after chemoradiation and into efficacy of new radiation techniques and swallowing and speech rehabilitation programs.

  5. Neural Spike-Train Analyses of the Speech-Based Envelope Power Spectrum Model

    PubMed Central

    Rallapalli, Varsha H.

    2016-01-01

    Diagnosing and treating hearing impairment is challenging because people with similar degrees of sensorineural hearing loss (SNHL) often have different speech-recognition abilities. The speech-based envelope power spectrum model (sEPSM) has demonstrated that the signal-to-noise ratio (SNRENV) from a modulation filter bank provides a robust speech-intelligibility measure across a wider range of degraded conditions than many long-standing models. In the sEPSM, noise (N) is assumed to: (a) reduce S + N envelope power by filling in dips within clean speech (S) and (b) introduce an envelope noise floor from intrinsic fluctuations in the noise itself. While the promise of SNRENV has been demonstrated for normal-hearing listeners, it has not been thoroughly extended to hearing-impaired listeners because of limited physiological knowledge of how SNHL affects speech-in-noise envelope coding relative to noise alone. Here, envelope coding to speech-in-noise stimuli was quantified from auditory-nerve model spike trains using shuffled correlograms, which were analyzed in the modulation-frequency domain to compute modulation-band estimates of neural SNRENV. Preliminary spike-train analyses show strong similarities to the sEPSM, demonstrating feasibility of neural SNRENV computations. Results suggest that individual differences can occur based on differential degrees of outer- and inner-hair-cell dysfunction in listeners currently diagnosed into the single audiological SNHL category. The predicted acoustic-SNR dependence in individual differences suggests that the SNR-dependent rate of susceptibility could be an important metric in diagnosing individual differences. Future measurements of the neural SNRENV in animal studies with various forms of SNHL will provide valuable insight for understanding individual differences in speech-in-noise intelligibility.

  6. Effect of dental wear, stabilization appliance and anterior tooth reconstruction on mandibular movements during speech.

    PubMed

    Serrano, Priscila de Oliveira; Faot, Fernanda; Del Bel Cury, Altair Antoninha; Rodrigues Garcia, Renata Cunha Matheus

    2008-01-01

    This study described changes in mandibular movements during pronunciation of /m/ and /s/ sounds in Portuguese, in patients presenting dental wear before and after appliance insertion and tooth reconstruction. Subjects were divided into a control group of dentate patients and an experimental group of patients with incisal tooth wear due to bruxism. A magnetic jaw tracking device measured the jaw opening, and translations to left and right sides of the mandible during pronunciation of phonemes. Evaluations were carried out 1 week and immediately before appliance insertion; 24 h, 7, 30 and 60 days after appliance insertion; and 1 week and 1 month after tooth reconstruction. Data were submitted to two-way ANOVA, Mann-Whitney and Friedman tests (p<0.05). Jaw opening was different (p<0.05) for both sounds in all periods. The anteroposterior amplitude for /s/ showed differences immediately before and 1 month after appliance insertion (p<0.05). Lateral amplitude for the right side showed differences between groups after appliance insertion for /s/, and 1 and 2 months after appliance insertion for the /m/ (p<0.05). Volunteers with anterior tooth wear had a wider opening movement, and the movements during speech of /m/ and /s/ sounds were not changed after appliance insertion and reconstruction of teeth.

  7. Non-right handed primary progressive apraxia of speech.

    PubMed

    Botha, Hugo; Duffy, Joseph R; Whitwell, Jennifer L; Strand, Edythe A; Machulda, Mary M; Spychalla, Anthony J; Tosakulwong, Nirubol; Senjem, Matthew L; Knopman, David S; Petersen, Ronald C; Jack, Clifford R; Lowe, Val J; Josephs, Keith A

    2018-07-15

    In recent years a large and growing body of research has greatly advanced our understanding of primary progressive apraxia of speech. Handedness has emerged as one potential marker of selective vulnerability in degenerative diseases. This study evaluated the clinical and imaging findings in non-right handed compared to right handed participants in a prospective cohort diagnosed with primary progressive apraxia of speech. A total of 30 participants were included. Compared to the expected rate in the population, there was a higher prevalence of non-right handedness among those with primary progressive apraxia of speech (6/30, 20%). Small group numbers meant that these results did not reach statistical significance, although the effect sizes were moderate-to-large. There were no clinical differences between right handed and non-right handed participants. Bilateral hypometabolism was seen in primary progressive apraxia of speech compared to controls, with non-right handed participants showing more right hemispheric involvement. This is the first report of a higher rate of non-right handedness in participants with isolated apraxia of speech, which may point to an increased vulnerability for developing this disorder among non-right handed participants. This challenges prior hypotheses about a relative protective effect of non-right handedness for tau-related neurodegeneration. We discuss potential avenues for future research to investigate the relationship between handedness and motor disorders more generally. Copyright © 2018 Elsevier B.V. All rights reserved.

  8. Effectiveness of Early Phonological Awareness Interventions for Students with Speech or Language Impairments

    PubMed Central

    Al Otaiba, Stephanie; Puranik, Cynthia; Zilkowski, Robin; Curran, Tricia

    2009-01-01

    This article reviews research examining the efficacy of early phonological interventions for young students identified with Speech or Language impairments. Eighteen studies are included, providing results for nearly 500 students in preschool through third grade. Although findings were generally positive, there were large individual differences in response to intervention. Further, there was little evidence that interventions enabled students to catch up in phonological or reading skills to typically developing peers. Methodological issues are described and implications for practice and future research are discussed. PMID:20161557

  9. Environmental Consciousness: Our Greatest Environmental Achievement.

    ERIC Educational Resources Information Center

    Hickel, Walter J.

    1991-01-01

    The text of a speech delivered on Earth Day, 1970 is presented. The significance of the event is discussed. Examples of progress made since 1970 are described and suggestions for the future provided. (CW)

  10. Early Intervention Programs in Bangladesh.

    ERIC Educational Resources Information Center

    Lowe, Armin

    1993-01-01

    This paper discusses the incidence of hearing impairment in Bangladesh, the struggle to achieve appropriate services for this population, the establishment of the National Centre for Hearing and Speech of Children, and future plans. (JDD)

  11. A Systematic Review of Tablet Computers and Portable Media Players as Speech Generating Devices for Individuals with Autism Spectrum Disorder.

    PubMed

    Lorah, Elizabeth R; Parnell, Ashley; Whitby, Peggy Schaefer; Hantula, Donald

    2015-12-01

    Powerful, portable, off-the-shelf handheld devices, such as tablet based computers (i.e., iPad(®); Galaxy(®)) or portable multimedia players (i.e., iPod(®)), can be adapted to function as speech generating devices for individuals with autism spectrum disorders or related developmental disabilities. This paper reviews the research in this new and rapidly growing area and delineates an agenda for future investigations. In general, participants using these devices acquired verbal repertoires quickly. Studies comparing these devices to picture exchange or manual sign language found that acquisition was often quicker when using a tablet computer and that the vast majority of participants preferred using the device to picture exchange or manual sign language. Future research in interface design, user experience, and extended verbal repertoires is recommended.

  12. Adopting public health approaches to communication disability: challenges for the education of speech-language pathologists.

    PubMed

    Wylie, Karen; McAllister, Lindy; Davidson, Bronwyn; Marshall, Julie; Law, James

    2014-01-01

    Public health approaches to communication disability challenge the profession of speech-language pathology (SLP) to reconsider both frames of reference for practice and models of education. This paper reviews the impetus for public health approaches to communication disability and considers how public health is, and could be, incorporated into SLP education, both now and in the future. The paper describes tensions between clinical services, which have become increasingly specialized, and public health approaches that offer a broader view of communication disability and communication disability prevention. It presents a discussion of these tensions and asserts that public health approaches to communication are themselves a specialist field, requiring specific knowledge and skills. The authors suggest the use of the term 'communication disability public health' to refer to this type of work and offer a preliminary definition in order to advance discussion. Examples from three countries are provided of how some SLP degree programmes are integrating public health into the SLP curriculum. Alternative models of training for communication disability public health that may be relevant in the future in different contexts and countries are presented, prompting the SLP profession to consider whether communication disability public health is a field of practice for speech-language pathologists or whether it has broader workforce implications. The paper concludes with some suggestions for the future which may advance thinking, research and practice in communication disability public health. © 2015 S. Karger AG, Basel.

  13. Speech recognition and parent-ratings from auditory development questionnaires in children who are hard of hearing

    PubMed Central

    McCreery, Ryan W.; Walker, Elizabeth A.; Spratford, Meredith; Oleson, Jacob; Bentler, Ruth; Holte, Lenore; Roush, Patricia

    2015-01-01

    Objectives Progress has been made in recent years in the provision of amplification and early intervention for children who are hard of hearing. However, children who use hearing aids (HA) may have inconsistent access to their auditory environment due to limitations in speech audibility through their HAs or limited HA use. The effects of variability in children’s auditory experience on parent-report auditory skills questionnaires and on speech recognition in quiet and in noise were examined for a large group of children who were followed as part of the Outcomes of Children with Hearing Loss study. Design Parent ratings on auditory development questionnaires and children’s speech recognition were assessed for 306 children who are hard of hearing. Children ranged in age from 12 months to 9 years of age. Three questionnaires involving parent ratings of auditory skill development and behavior were used, including the LittlEARS Auditory Questionnaire, Parents Evaluation of Oral/Aural Performance in Children Rating Scale, and an adaptation of the Speech, Spatial and Qualities of Hearing scale. Speech recognition in quiet was assessed using the Open and Closed set task, Early Speech Perception Test, Lexical Neighborhood Test, and Phonetically-balanced Kindergarten word lists. Speech recognition in noise was assessed using the Computer-Assisted Speech Perception Assessment. Children who are hard of hearing were compared to peers with normal hearing matched for age, maternal educational level and nonverbal intelligence. The effects of aided audibility, HA use and language ability on parent responses to auditory development questionnaires and on children’s speech recognition were also examined. Results Children who are hard of hearing had poorer performance than peers with normal hearing on parent ratings of auditory skills and had poorer speech recognition. Significant individual variability among children who are hard of hearing was observed. Children with greater aided audibility through their HAs, more hours of HA use and better language abilities generally had higher parent ratings of auditory skills and better speech recognition abilities in quiet and in noise than peers with less audibility, more limited HA use or poorer language abilities. In addition to the auditory and language factors that were predictive for speech recognition in quiet, phonological working memory was also a positive predictor for word recognition abilities in noise. Conclusions Children who are hard of hearing continue to experience delays in auditory skill development and speech recognition abilities compared to peers with normal hearing. However, significant improvements in these domains have occurred in comparison to similar data reported prior to the adoption of universal newborn hearing screening and early intervention programs for children who are hard of hearing. Increasing the audibility of speech has a direct positive effect on auditory skill development and speech recognition abilities, and may also enhance these skills by improving language abilities in children who are hard of hearing. Greater number of hours of HA use also had a significant positive impact on parent ratings of auditory skills and children’s speech recognition. PMID:26731160

  14. Improvement of a Vocal Fold Imaging System

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Krauter, K. G.

    Medical professionals can better serve their patients through continual update of their imaging tools. A wide range of pathologies and disease may afflict human vocal cords or, as they’re also known, vocal folds. These diseases can affect human speech hampering the ability of the patient to communicate. Vocal folds must be opened for breathing and the closed to produce speech. Currently methodologies to image markers of potential pathologies are difficult to use and often fail to detect early signs of disease. These current methodologies rely on a strobe light and slower frame rate camera in an attempt to obtain imagesmore » as the vocal folds travel over the full extent of their motion.« less

  15. Musical intervention enhances infants’ neural processing of temporal structure in music and speech

    PubMed Central

    Zhao, T. Christina; Kuhl, Patricia K.

    2016-01-01

    Individuals with music training in early childhood show enhanced processing of musical sounds, an effect that generalizes to speech processing. However, the conclusions drawn from previous studies are limited due to the possible confounds of predisposition and other factors affecting musicians and nonmusicians. We used a randomized design to test the effects of a laboratory-controlled music intervention on young infants’ neural processing of music and speech. Nine-month-old infants were randomly assigned to music (intervention) or play (control) activities for 12 sessions. The intervention targeted temporal structure learning using triple meter in music (e.g., waltz), which is difficult for infants, and it incorporated key characteristics of typical infant music classes to maximize learning (e.g., multimodal, social, and repetitive experiences). Controls had similar multimodal, social, repetitive play, but without music. Upon completion, infants’ neural processing of temporal structure was tested in both music (tones in triple meter) and speech (foreign syllable structure). Infants’ neural processing was quantified by the mismatch response (MMR) measured with a traditional oddball paradigm using magnetoencephalography (MEG). The intervention group exhibited significantly larger MMRs in response to music temporal structure violations in both auditory and prefrontal cortical regions. Identical results were obtained for temporal structure changes in speech. The intervention thus enhanced temporal structure processing not only in music, but also in speech, at 9 mo of age. We argue that the intervention enhanced infants’ ability to extract temporal structure information and to predict future events in time, a skill affecting both music and speech processing. PMID:27114512

  16. Musical intervention enhances infants' neural processing of temporal structure in music and speech.

    PubMed

    Zhao, T Christina; Kuhl, Patricia K

    2016-05-10

    Individuals with music training in early childhood show enhanced processing of musical sounds, an effect that generalizes to speech processing. However, the conclusions drawn from previous studies are limited due to the possible confounds of predisposition and other factors affecting musicians and nonmusicians. We used a randomized design to test the effects of a laboratory-controlled music intervention on young infants' neural processing of music and speech. Nine-month-old infants were randomly assigned to music (intervention) or play (control) activities for 12 sessions. The intervention targeted temporal structure learning using triple meter in music (e.g., waltz), which is difficult for infants, and it incorporated key characteristics of typical infant music classes to maximize learning (e.g., multimodal, social, and repetitive experiences). Controls had similar multimodal, social, repetitive play, but without music. Upon completion, infants' neural processing of temporal structure was tested in both music (tones in triple meter) and speech (foreign syllable structure). Infants' neural processing was quantified by the mismatch response (MMR) measured with a traditional oddball paradigm using magnetoencephalography (MEG). The intervention group exhibited significantly larger MMRs in response to music temporal structure violations in both auditory and prefrontal cortical regions. Identical results were obtained for temporal structure changes in speech. The intervention thus enhanced temporal structure processing not only in music, but also in speech, at 9 mo of age. We argue that the intervention enhanced infants' ability to extract temporal structure information and to predict future events in time, a skill affecting both music and speech processing.

  17. Robust speech perception: Recognize the familiar, generalize to the similar, and adapt to the novel

    PubMed Central

    Kleinschmidt, Dave F.; Jaeger, T. Florian

    2016-01-01

    Successful speech perception requires that listeners map the acoustic signal to linguistic categories. These mappings are not only probabilistic, but change depending on the situation. For example, one talker’s /p/ might be physically indistinguishable from another talker’s /b/ (cf. lack of invariance). We characterize the computational problem posed by such a subjectively non-stationary world and propose that the speech perception system overcomes this challenge by (1) recognizing previously encountered situations, (2) generalizing to other situations based on previous similar experience, and (3) adapting to novel situations. We formalize this proposal in the ideal adapter framework: (1) to (3) can be understood as inference under uncertainty about the appropriate generative model for the current talker, thereby facilitating robust speech perception despite the lack of invariance. We focus on two critical aspects of the ideal adapter. First, in situations that clearly deviate from previous experience, listeners need to adapt. We develop a distributional (belief-updating) learning model of incremental adaptation. The model provides a good fit against known and novel phonetic adaptation data, including perceptual recalibration and selective adaptation. Second, robust speech recognition requires listeners learn to represent the structured component of cross-situation variability in the speech signal. We discuss how these two aspects of the ideal adapter provide a unifying explanation for adaptation, talker-specificity, and generalization across talkers and groups of talkers (e.g., accents and dialects). The ideal adapter provides a guiding framework for future investigations into speech perception and adaptation, and more broadly language comprehension. PMID:25844873

  18. Speech Discrimination Difficulties in High-Functioning Autism Spectrum Disorder Are Likely Independent of Auditory Hypersensitivity

    PubMed Central

    Dunlop, William A.; Enticott, Peter G.; Rajan, Ramesh

    2016-01-01

    Autism Spectrum Disorder (ASD), characterized by impaired communication skills and repetitive behaviors, can also result in differences in sensory perception. Individuals with ASD often perform normally in simple auditory tasks but poorly compared to typically developed (TD) individuals on complex auditory tasks like discriminating speech from complex background noise. A common trait of individuals with ASD is hypersensitivity to auditory stimulation. No studies to our knowledge consider whether hypersensitivity to sounds is related to differences in speech-in-noise discrimination. We provide novel evidence that individuals with high-functioning ASD show poor performance compared to TD individuals in a speech-in-noise discrimination task with an attentionally demanding background noise, but not in a purely energetic noise. Further, we demonstrate in our small sample that speech-hypersensitivity does not appear to predict performance in the speech-in-noise task. The findings support the argument that an attentional deficit, rather than a perceptual deficit, affects the ability of individuals with ASD to discriminate speech from background noise. Finally, we piloted a novel questionnaire that measures difficulty hearing in noisy environments, and sensitivity to non-verbal and verbal sounds. Psychometric analysis using 128 TD participants provided novel evidence for a difference in sensitivity to non-verbal and verbal sounds, and these findings were reinforced by participants with ASD who also completed the questionnaire. The study was limited by a small and high-functioning sample of participants with ASD. Future work could test larger sample sizes and include lower-functioning ASD participants. PMID:27555814

  19. Asthma, hay fever, and food allergy are associated with caregiver-reported speech disorders in US children

    PubMed Central

    Strom, Mark A.; Silverberg, Jonathan I.

    2016-01-01

    Background Children with asthma, hay fever, and food allergy may have several factors that increase their risk of speech disorder, including allergic inflammation, ADD/ADHD, and sleep disturbance. However, few studies have examined a relationship between asthma, allergic disease, and speech disorder. We sought to determine whether asthma, hay fever, and food allergy are associated with speech disorder in children and whether disease severity, sleep disturbance, or ADD/ADHD modified such associations. Methods We analyzed cross-sectional data on 337,285 children aged 2–17 years from 19 US population-based studies, including the 1997–2013 National Health Interview Survey and the 2003/4 and 2007/8 National Survey of Children’s Health. Results In multivariate models, controlling for age, demographic factors, healthcare utilization, and history of eczema, lifetime history of asthma (odds ratio [95% confidence interval]: 1.18 [1.04–1.34], p = 0.01), and one-year history of hay fever (1.44 [1.28–1.62], p < 0.0001) and food allergy (1.35 [1.13–1.62], p = 0.001) were associated with increased odds of speech disorder. Children with current (1.37 [1.15–1.59] p = 0.0003) but not past (p = 0.06) asthma had increased risk of speech disorder. In one study that assessed caregiver-reported asthma severity, mild (1.58 [1.20–2.08], p = 0.001) and moderate (2.99 [1.54–3.41], p < 0.0001) asthma were associated with increased odds of speech disorder; however, severe asthma was associated with the highest odds of speech disorder (5.70 [2.36–13.78], p = 0.0001). Conclusion Childhood asthma, hay fever, and food allergy are associated with increased risk of speech disorder. Future prospective studies are needed to characterize the associations. PMID:27091599

  20. Asthma, hay fever, and food allergy are associated with caregiver-reported speech disorders in US children.

    PubMed

    Strom, Mark A; Silverberg, Jonathan I

    2016-09-01

    Children with asthma, hay fever, and food allergy may have several factors that increase their risk of speech disorder, including allergic inflammation, ADD/ADHD, and sleep disturbance. However, few studies have examined a relationship between asthma, allergic disease, and speech disorder. We sought to determine whether asthma, hay fever, and food allergy are associated with speech disorder in children and whether disease severity, sleep disturbance, or ADD/ADHD modified such associations. We analyzed cross-sectional data on 337,285 children aged 2-17 years from 19 US population-based studies, including the 1997-2013 National Health Interview Survey and the 2003/4 and 2007/8 National Survey of Children's Health. In multivariate models, controlling for age, demographic factors, healthcare utilization, and history of eczema, lifetime history of asthma (odds ratio [95% confidence interval]: 1.18 [1.04-1.34], p = 0.01), and one-year history of hay fever (1.44 [1.28-1.62], p < 0.0001) and food allergy (1.35 [1.13-1.62], p = 0.001) were associated with increased odds of speech disorder. Children with current (1.37 [1.15-1.59] p = 0.0003) but not past (p = 0.06) asthma had increased risk of speech disorder. In one study that assessed caregiver-reported asthma severity, mild (1.58 [1.20-2.08], p = 0.001) and moderate (2.99 [1.54-3.41], p < 0.0001) asthma were associated with increased odds of speech disorder; however, severe asthma was associated with the highest odds of speech disorder (5.70 [2.36-13.78], p = 0.0001). Childhood asthma, hay fever, and food allergy are associated with increased risk of speech disorder. Future prospective studies are needed to characterize the associations. © 2016 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  1. The Personality of Past, Present and Future Speech-Language Pathology Students

    ERIC Educational Resources Information Center

    Byrne, Nicole

    2018-01-01

    Background: As allied health professions change over time to keep up with and reflect a rapidly changing society, it is quite possible that the people attracted to the profession may also change. If this is the case, then knowing this could be critical for future workforce marketing, training and planning. Aims: The aim was to investigate whether…

  2. The Future of Institutional Research. Proceedings of the Annual Conference of the Southern Association for Institutional Research (Orlando, Florida, October 24-26, 1979).

    ERIC Educational Resources Information Center

    Salley, Charles D., Ed.; And Others

    Proceedings of the 1979 conference of the Southern Association for Institutional Research (SAIR) are presented. The conference theme was the future of institutional research. Contents include reports of preconference workshops, speeches, panel reports, abstracts of papers, and reports pertaining to the affairs of the SAIR. Documents and authors…

  3. Letters from George Washington and Samuel Cabble, and Speeches by Franklin D. Roosevelt and John F. Kennedy

    ERIC Educational Resources Information Center

    Potter, Lee Ann

    2008-01-01

    In this article, the author uses several primary sources to demonstrate that George Washington, Samuel Cabble, Franklin D. Roosevelt, and John F. Kennedy stated their awareness of contemporary challenges, but looked to the future with hope and optimism. When they envisioned the future, their words indicated that they did not just imagine it, but…

  4. Sensory-Cognitive Interaction in the Neural Encoding of Speech in Noise: A Review

    PubMed Central

    Anderson, Samira; Kraus, Nina

    2011-01-01

    Background Speech-in-noise (SIN) perception is one of the most complex tasks faced by listeners on a daily basis. Although listening in noise presents challenges for all listeners, background noise inordinately affects speech perception in older adults and in children with learning disabilities. Hearing thresholds are an important factor in SIN perception, but they are not the only factor. For successful comprehension, the listener must perceive and attend to relevant speech features, such as the pitch, timing, and timbre of the target speaker’s voice. Here, we review recent studies linking SIN and brainstem processing of speech sounds. Purpose To review recent work that has examined the ability of the auditory brainstem response to complex sounds (cABR), which reflects the nervous system’s transcription of pitch, timing, and timbre, to be used as an objective neural index for hearing-in-noise abilities. Study Sample We examined speech-evoked brainstem responses in a variety of populations, including children who are typically developing, children with language-based learning impairment, young adults, older adults, and auditory experts (i.e., musicians). Data Collection and Analysis In a number of studies, we recorded brainstem responses in quiet and babble noise conditions to the speech syllable /da/ in all age groups, as well as in a variable condition in children in which /da/ was presented in the context of seven other speech sounds. We also measured speech-in-noise perception using the Hearing-in-Noise Test (HINT) and the Quick Speech-in-Noise Test (QuickSIN). Results Children and adults with poor SIN perception have deficits in the subcortical spectrotemporal representation of speech, including low-frequency spectral magnitudes and the timing of transient response peaks. Furthermore, auditory expertise, as engendered by musical training, provides both behavioral and neural advantages for processing speech in noise. Conclusions These results have implications for future assessment and management strategies for young and old populations whose primary complaint is difficulty hearing in background noise. The cABR provides a clinically applicable metric for objective assessment of individuals with SIN deficits, for determination of the biologic nature of disorders affecting SIN perception, for evaluation of appropriate hearing aid algorithms, and for monitoring the efficacy of auditory remediation and training. PMID:21241645

  5. Initial Progress Toward Development of a Voice-Based Computer-Delivered Motivational Intervention for Heavy Drinking College Students: An Experimental Study.

    PubMed

    Kahler, Christopher W; Lechner, William J; MacGlashan, James; Wray, Tyler B; Littman, Michael L

    2017-06-28

    Computer-delivered interventions have been shown to be effective in reducing alcohol consumption in heavy drinking college students. However, these computer-delivered interventions rely on mouse, keyboard, or touchscreen responses for interactions between the users and the computer-delivered intervention. The principles of motivational interviewing suggest that in-person interventions may be effective, in part, because they encourage individuals to think through and speak aloud their motivations for changing a health behavior, which current computer-delivered interventions do not allow. The objective of this study was to take the initial steps toward development of a voice-based computer-delivered intervention that can ask open-ended questions and respond appropriately to users' verbal responses, more closely mirroring a human-delivered motivational intervention. We developed (1) a voice-based computer-delivered intervention that was run by a human controller and that allowed participants to speak their responses to scripted prompts delivered by speech generation software and (2) a text-based computer-delivered intervention that relied on the mouse, keyboard, and computer screen for all interactions. We randomized 60 heavy drinking college students to interact with the voice-based computer-delivered intervention and 30 to interact with the text-based computer-delivered intervention and compared their ratings of the systems as well as their motivation to change drinking and their drinking behavior at 1-month follow-up. Participants reported that the voice-based computer-delivered intervention engaged positively with them in the session and delivered content in a manner consistent with motivational interviewing principles. At 1-month follow-up, participants in the voice-based computer-delivered intervention condition reported significant decreases in quantity, frequency, and problems associated with drinking, and increased perceived importance of changing drinking behaviors. In comparison to the text-based computer-delivered intervention condition, those assigned to voice-based computer-delivered intervention reported significantly fewer alcohol-related problems at the 1-month follow-up (incident rate ratio 0.60, 95% CI 0.44-0.83, P=.002). The conditions did not differ significantly on perceived importance of changing drinking or on measures of drinking quantity and frequency of heavy drinking. Results indicate that it is feasible to construct a series of open-ended questions and a bank of responses and follow-up prompts that can be used in a future fully automated voice-based computer-delivered intervention that may mirror more closely human-delivered motivational interventions to reduce drinking. Such efforts will require using advanced speech recognition capabilities and machine-learning approaches to train a program to mirror the decisions made by human controllers in the voice-based computer-delivered intervention used in this study. In addition, future studies should examine enhancements that can increase the perceived warmth and empathy of voice-based computer-delivered intervention, possibly through greater personalization, improvements in the speech generation software, and embodying the computer-delivered intervention in a physical form. ©Christopher W Kahler, William J Lechner, James MacGlashan, Tyler B Wray, Michael L Littman. Originally published in JMIR Mental Health (http://mental.jmir.org), 28.06.2017.

  6. Vocal Control: Is It Susceptible to the Negative Effects of Self-Regulatory Depletion?

    PubMed

    Vinney, Lisa A; van Mersbergen, Miriam; Connor, Nadine P; Turkstra, Lyn S

    2016-09-01

    Self-regulation (SR) relies on the capacity to modify behavior. This capacity may diminish with use and result in self-regulatory depletion (SRD), or the reduced ability to engage in future SR efforts. If the SRD effect applies to vocal behavior, it may hinder success during behavioral voice treatment. Thus, this proof-of-concept study sought to determine whether SRD affects vocal behavior change and if so, whether it can be repaired by an intervention meant to replete SR resources. One hundred four women without voice disorders were randomized into groups that performed either (1) a high-SR writing task followed by a high-SR voice task; (2) a low-SR writing task followed by a high-SR voice task; or (3) a high-SR writing task followed by a relaxation intervention and a high-SR voice task. The high-SR voice tasks in all groups involved suppression of the Lombard effect during reading and free speech. The low-SR group suppressed the Lombard effect to a greater extent than the high-SR group and high-SR-plus-relaxation group on the free speech task. There were no significant group differences on the reading task. Findings suggest that SRD may present challenges to vocal behavior modification during free speech but not reading. Furthermore, relaxation did not significantly replete self-regulatory resources for vocal modification during free speech. Findings may highlight potential considerations for voice treatment and assessment and support the need for future research focusing on effective methods to test self-regulatory capacity and replete self-regulatory resources in voice patients. Published by Elsevier Inc.

  7. Cochlear implants: a remarkable past and a brilliant future

    PubMed Central

    Wilson, Blake S.; Dorman, Michael F.

    2013-01-01

    The aims of this paper are to (i) provide a brief history of cochlear implants; (ii) present a status report on the current state of implant engineering and the levels of speech understanding enabled by that engineering; (iii) describe limitations of current signal processing strategies and (iv) suggest new directions for research. With current technology the “average” implant patient, when listening to predictable conversations in quiet, is able to communicate with relative ease. However, in an environment typical of a workplace the average patient has a great deal of difficulty. Patients who are “above average” in terms of speech understanding, can achieve 100% correct scores on the most difficult tests of speech understanding in quiet but also have significant difficulty when signals are presented in noise. The major factors in these outcomes appear to be (i) a loss of low-frequency, fine structure information possibly due to the envelope extraction algorithms common to cochlear implant signal processing; (ii) a limitation in the number of effective channels of stimulation due to overlap in electric fields from electrodes, and (iii) central processing deficits, especially for patients with poor speech understanding. Two recent developments, bilateral implants and combined electric and acoustic stimulation, have promise to remediate some of the difficulties experienced by patients in noise and to reinstate low-frequency fine structure information. If other possibilities are realized, e.g., electrodes that emit drugs to inhibit cell death following trauma and to induce the growth of neurites toward electrodes, then the future is very bright indeed. PMID:18616994

  8. "The Seventh Seal."

    ERIC Educational Resources Information Center

    Palmer, Peter M.

    1969-01-01

    The significance of Bergman's "Seventh Seal" lies not in the speeches nor in the actions of the central characters but rather in the film's form, its totality created by the emotive elements of imagery and sound together with the intellectual elements of actions and words. The scene-units are related to a central motif (the opening of…

  9. Ironic Openings: The Interpretive Challenge of the "Black Manifesto"

    ERIC Educational Resources Information Center

    Parker, Maegan

    2008-01-01

    For generations, critics have dismissed James Forman's "Black Manifesto" as a rhetorical failure. Such judgments tend to focus on the prophetic and retributive registers of the speech and fail to account for the full range of its ironic structuration. By examining the complex interchange of prophetic, retributive, and tragic registers through…

  10. "Universal Responsiveness" or "Splendid Isolation?" Episodes from the History of Mathematics Education in Russia

    ERIC Educational Resources Information Center

    Karp, Alexander

    2006-01-01

    This article investigates the prevalent attitudes toward foreign influences and methodologies in Russian mathematics education at different periods in Russian history. The words "universal responsiveness" belong to Dostoevsky, who, in his famous speech on Pushkin, used them to characterize Pushkin's openness to the genius of all other…

  11. Barriers to Caregiver Compliance with Eating and Drinking Recommendations for Adults with Intellectual Disabilities and Dysphagia

    ERIC Educational Resources Information Center

    Chadwick, Darren D.; Jolliffe, Jane; Goldbart, Juliet; Burton, Mark H.

    2006-01-01

    Background: There is scant research on the subject of dysphagia and people with intellectual disabilities. This study explores the barriers which caregivers believe make following Speech and Language Therapists' (SLTs) dysphagia management strategies more difficult. Method: Semi-structured open-ended interviews were conducted with 46 caregivers…

  12. Electronic Field Trip: Journalism's New Frontier Involves Live, Interactive Broadcast.

    ERIC Educational Resources Information Center

    LaMar, Jason

    1998-01-01

    Describes the "Newseum," a recently opened museum in Arlington, Virginia, dedicated to journalism and freedom of speech. Lists its highlights: free admission, an interactive newsroom, a 126-foot video wall, a news history gallery, a domed theater with a 20-by-40-foot high-definition video screen, and "The Freedom Wall." (PA)

  13. "Changes in Purchasing Practices of School Insurances -- or, To Bid or not To Bid."

    ERIC Educational Resources Information Center

    Crockett, James E.

    This speech concentrates on a discussion of the increasing difficulty that school districts are experiencing in obtaining insurance coverage. The speaker offers school administrators some suggestions to assist them in securing the necessary coverage and suggests that when an insurance market is open and competitive, purchasing through bidding is…

  14. Automatic Classification of Question & Answer Discourse Segments from Teacher's Speech in Classrooms

    ERIC Educational Resources Information Center

    Blanchard, Nathaniel; D'Mello, Sidney; Olney, Andrew M.; Nystrand, Martin

    2015-01-01

    Question-answer (Q&A) is fundamental for dialogic instruction, an important pedagogical technique based on the free exchange of ideas and open-ended discussion. Automatically detecting Q&A is key to providing teachers with feedback on appropriate use of dialogic instructional strategies. In line with this, this paper studies the…

  15. Workplace Violence, Hate Crime and Free Speech: A Proactive Approach.

    ERIC Educational Resources Information Center

    Romas, Ted; Parmer, Harry

    For community colleges, crime on campus is a complex and pressing issue since most maintain open campuses with no physical security features, such as entry control points or fencing. Workplace crimes can take the form of robbery or other violent crimes, domestic and misdirected affection cases, employer-directed situations, terrorism, and hate…

  16. Open Communication: Having Your Voice Heard

    ERIC Educational Resources Information Center

    Brown, Angela

    2015-01-01

    Communication is the use of language to exchange information to one another. African slaves used to embark on communication by means of using common symbols and speech, telling stories, singing spirituals, writing poems. As time revolved, blacks valued education. Education and the ability to read write and effectively would give them the skill or…

  17. Sound Source Localization and Speech Understanding in Complex Listening Environments by Single-sided Deaf Listeners After Cochlear Implantation.

    PubMed

    Zeitler, Daniel M; Dorman, Michael F; Natale, Sarah J; Loiselle, Louise; Yost, William A; Gifford, Rene H

    2015-09-01

    To assess improvements in sound source localization and speech understanding in complex listening environments after unilateral cochlear implantation for single-sided deafness (SSD). Nonrandomized, open, prospective case series. Tertiary referral center. Nine subjects with a unilateral cochlear implant (CI) for SSD (SSD-CI) were tested. Reference groups for the task of sound source localization included young (n = 45) and older (n = 12) normal-hearing (NH) subjects and 27 bilateral CI (BCI) subjects. Unilateral cochlear implantation. Sound source localization was tested with 13 loudspeakers in a 180 arc in front of the subject. Speech understanding was tested with the subject seated in an 8-loudspeaker sound system arrayed in a 360-degree pattern. Directionally appropriate noise, originally recorded in a restaurant, was played from each loudspeaker. Speech understanding in noise was tested using the Azbio sentence test and sound source localization quantified using root mean square error. All CI subjects showed poorer-than-normal sound source localization. SSD-CI subjects showed a bimodal distribution of scores: six subjects had scores near the mean of those obtained by BCI subjects, whereas three had scores just outside the 95th percentile of NH listeners. Speech understanding improved significantly in the restaurant environment when the signal was presented to the side of the CI. Cochlear implantation for SSD can offer improved speech understanding in complex listening environments and improved sound source localization in both children and adults. On tasks of sound source localization, SSD-CI patients typically perform as well as BCI patients and, in some cases, achieve scores at the upper boundary of normal performance.

  18. Cochlear implant rehabilitation outcomes in Waardenburg syndrome children.

    PubMed

    de Sousa Andrade, Susana Margarida; Monteiro, Ana Rita Tomé; Martins, Jorge Humberto Ferreira; Alves, Marisa Costa; Santos Silva, Luis Filipe; Quadros, Jorge Manuel Cardoso; Ribeiro, Carlos Alberto Reis

    2012-09-01

    The purpose of this study was to review the outcomes of children with documented Waardenburg syndrome implanted in the ENT Department of Centro Hospitalar de Coimbra, concerning postoperative speech perception and production, in comparison to the rest of non-syndromic implanted children. A retrospective chart review was performed for children congenitally deaf who had undergone cochlear implantation with multichannel implants, diagnosed as having Waardenburg syndrome, between 1992 and 2011. Postoperative performance outcomes were assessed and confronted with results obtained by children with non-syndromic congenital deafness also implanted in our department. Open-set auditory perception skills were evaluated by using European Portuguese speech discrimination tests (vowels test, monosyllabic word test, number word test and words in sentence test). Meaningful auditory integration scales (MAIS) and categories of auditory performance (CAP) were also measured. Speech production was further assessed and included results on meaningful use of speech Scale (MUSS) and speech intelligibility rating (SIR). To date, 6 implanted children were clinically identified as having WS type I, and one met the diagnosis of type II. All WS children received multichannel cochlear implants, with a mean age at implantation of 30.6±9.7months (ranging from 19 to 42months). Postoperative outcomes in WS children were similar to other nonsyndromic children. In addition, in number word and vowels discrimination test WS group showed slightly better performances, as well as in MUSS and MAIS assessment. Our study has shown that cochlear implantation should be considered a rehabilitative option for Waardenburg syndrome children with profound deafness, enabling the development and improvement of speech perception and production abilities in this group of patients, reinforcing their candidacy for this audio-oral rehabilitation method. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  19. Non-speech oro-motor exercise use in acquired dysarthria management: regimes and rationales.

    PubMed

    Mackenzie, Catherine; Muir, Margaret; Allen, Carolyn

    2010-01-01

    Non-speech oro-motor exercises (NSOMExs) are described in speech and language therapy manuals and are thought to be much used in acquired dysarthria intervention, though there is no robust evidence of an influence on speech outcome. Opinions differ as to whether, and for which dysarthria presentations, NSOMExs are appropriate. The investigation sought to collect development-phase data, in accordance with the Medical Research Council (MRC) evaluation of complex interventions. The aims were to establish the extent of use of NSOMExs in acquired disorders, the exercise regimes in use for dysarthria, with which dysarthric populations, and the anticipated clinical outcomes. A further aim was to determine the influencing rationales where NSOMExs were or were not used in dysarthria intervention. Speech and language therapists throughout Scotland, Wales, and Northern Ireland, working with adult-acquired dysarthria, were identified by their service heads. They received postal questionnaires comprising 21 closed and two open questions, covering respondent biographics, use of NSOMExs, anticipated clinical outcomes, and practice influencing rationales. One hundred and ninety-one (56% response) completed questionnaires were returned. Eighty-one per cent of respondents used NSOMExs in dysarthria. There was no association with years of speech and language therapy experience. Those who used and those who did not use NSOMExs provided similar influencing rationales, including evidence from their own practice, and Higher Education Institute teaching. More experienced speech and language therapists were more likely than those more recently qualified to be guided by results from their own practice. Input from the attended Higher Education Institute was more influential for those less experienced than for those more experienced. Clinical outcome aims were not confined to speech, but also included improvements in movement, sensory awareness, appearance, emotional status, dysphagia and drooling. NSOMExs were used with many neurological disorders, especially stroke, all dysarthria classes, especially flaccid, and all severity levels. Tongue and lip exercises were more frequent than face, jaw and soft palate. The most common regimes were four to six repetitions of each exercise, during three practice periods daily, each of 6–10 min. NSOMExs are a frequent component of dysarthria management in the UK-devolved government countries. This confirmation, along with the details of speech and language therapy practice, provides a foundation for clinical research which will compare outcomes for people with dysarthria, whose management includes and does not include NSOMExs. Speech and language therapy practice may be guided by evidence that speech outcome is or is not affected by NSOMExs.

  20. [Acoustic conditions in open plan offices - Pilot test results].

    PubMed

    Mikulski, Witold

    The main source of noise in open plan office are conversations. Office work standards in such premises are attained by applying specific acoustic adaptation. This article presents the results of pilot tests and acoustic evaluation of open space rooms. Acoustic properties of 6 open plan office rooms were the subject of the tests. Evaluation parameters, measurement methods and criterial values were adopted according to the following standards: PN-EN ISO 3382- 3:2012, PN-EN ISO 3382-2:2010, PN-B-02151-4:2015-06 and PN-B-02151-3:2015-10. The reverberation time was 0.33- 0.55 s (maximum permissible value in offices - 0.6 s; the criterion was met), sound absorption coefficient in relation to 1 m2 of the room's plan was 0.77-1.58 m2 (minimum permissible value - 1.1 m2; 2 out of 6 rooms met the criterion), distraction distance was 8.5-14 m (maximum permissible value - 5 m; none of the rooms met the criterion), A-weighted sound pressure level of speech at a distance of 4 m was 43.8-54.7 dB (maximum permissible value - 48 dB; 2 out of 6 rooms met the criterion), spatial decay rate of the speech was 1.8-6.3 dB (minimum permissible value - 7 dB; none of the rooms met the criterion). Standard acoustic treatment, containing sound absorbing suspended ceiling, sound absorbing materials on the walls, carpet flooring and sound absorbing workplace barriers, is not sufficient. These rooms require specific advanced acoustic solutions. Med Pr 2016;67(5):653-662. This work is available in Open Access model and licensed under a CC BY-NC 3.0 PL license.

  1. How to become an authentic speaker. Even sincere speeches often come across as contrived. A four-step process will help you create a true emotional connection with your audience.

    PubMed

    Morgan, Nick

    2008-11-01

    Like the best-laid schemes of mice and men, the best-rehearsed speeches go oft astray. No amount of preparation can counter an audience's perception that the speaker is calculating or insincere. Why do so many managers have trouble communicating authenticity to their listeners? Morgan, a communications coach for more than two decades, offers advice for overcoming this difficulty. Recent brain research shows that natural, unstudied gestures--what Morgan calls the " second conversation"--express emotions or impulses a split second before our thought processes have turned them into words. So the timing of practiced gestures will always be subtly off--just enough to be picked up by listeners' unconscious ability to read body language. If you can't practice the unspoken part of your delivery, what can you do? Tap into four basic impulses underlying your speech--to be open to the audience, to connect with it, to be passionate, and to "listen" to how the audience is responding--and then rehearse your presentation with each in mind. You can become more open, for instance, by imagining that you're speaking to your spouse or close friend. To more readily connect, focus on needing to engage your listeners and then to keep their attention, as if you were speaking to a child who isn't heeding your words. To convey your passion, identify the feelings behind your speech and let them come through. To listen, think about what the audience is probably feeling when you step up to the podium and be alert to the nonverbal messages of its members. Internalizing these four impulses as you practice will help you come across as relaxed and authentic--your body language will take care of itself.

  2. The speech aversion hypothesis has explanatory power in a Minimal Speech Approach to aloof, non-verbal, severe autism.

    PubMed

    Whittaker, Christopher A

    2012-01-01

    In the search for 'pure autism', non-verbal children labeled aloof, Severely Autistic with Developmental Disabilities (ASA/DD), are routinely excluded from psychological research. This exclusion is predicated on the claim that they are indistinguishable from those with SLD/PMLD, which is refuted through a discussion of the extant literature. A novel, falsifiable, speech aversion hypothesis is proposed: "aloof, non-verbal young children (<7 years), with severe autism (CARS≥37), but without significant dysmorphic features, will show aversive reactions to complex speech (>2-3 words), but not to a silent interlocutor, or one imitating their vocalizations, in proximal encounters." Implications are examined by deconstructing the presenting symptoms of ASA/DD in response to the hypothesis. Supporting evidence is drawn from: Minimal Speech Approach (MSA) research showing high levels of spontaneous requests for social routines; a reinterpretation of still-face research as a still-(silent)-face paradigm; auditory processing MMN data employing EEG/MEG; and possible links to epileptiform activity and verbal auditory agnosia. Guidelines are established for future research. This hypothesis, if corroborated, would add to the auditory processing anomalies seen in severe autism and lead to synergies of existing and new areas of research, with significant theoretical, therapeutic, and educational implications. Copyright © 2011 Elsevier Ltd. All rights reserved.

  3. Imitative Production of Rising Speech Intonation in Pediatric Cochlear Implant Recipients

    PubMed Central

    Peng, Shu-Chen; Tomblin, J. Bruce; Spencer, Linda J.; Hurtig, Richard R.

    2011-01-01

    Purpose This study investigated the acoustic characteristics of pediatric cochlear implant (CI) recipients' imitative production of rising speech intonation, in relation to the perceptual judgments by listeners with normal hearing (NH). Method Recordings of a yes–no interrogative utterance imitated by 24 prelingually deafened children with a CI were extracted from annual evaluation sessions. These utterances were perceptually judged by adult NH listeners in regard with intonation contour type (non-rise, partial-rise, or full-rise) and contour appropriateness (on a 5-point scale). Fundamental frequency, intensity, and duration properties of each utterance were also acoustically analyzed. Results Adult NH listeners' judgments of intonation contour type and contour appropriateness for each CI participant 's utterances were highly positively correlated. The pediatric CI recipients did not consistently use appropriate intonation contours when imitating a yes–no question. Acoustic properties of speech intonation produced by these individuals were discernible among utterances of different intonation contour types according to NH listeners' perceptual judgments. Conclusions These findings delineated the perceptual and acoustic characteristics of speech intonation imitated by prelingually deafened children and young adults with a CI. Future studies should address whether the degraded signals these individuals perceive via a CI contribute to their difficulties with speech intonation production. PMID:17905907

  4. Binaural unmasking of multi-channel stimuli in bilateral cochlear implant users.

    PubMed

    Van Deun, Lieselot; van Wieringen, Astrid; Francart, Tom; Büchner, Andreas; Lenarz, Thomas; Wouters, Jan

    2011-10-01

    Previous work suggests that bilateral cochlear implant users are sensitive to interaural cues if experimental speech processors are used to preserve accurate interaural information in the electrical stimulation pattern. Binaural unmasking occurs in adults and children when an interaural delay is applied to the envelope of a high-rate pulse train. Nevertheless, for speech perception, binaural unmasking benefits have not been demonstrated consistently, even with coordinated stimulation at both ears. The present study aimed at bridging the gap between basic psychophysical performance on binaural signal detection tasks on the one hand and binaural perception of speech in noise on the other hand. Therefore, binaural signal detection was expanded to multi-channel stimulation and biologically relevant interaural delays. A harmonic complex, consisting of three sinusoids (125, 250, and 375 Hz), was added to three 125-Hz-wide noise bands centered on the sinusoids. When an interaural delay of 700 μs was introduced, an average BMLD of 3 dB was established. Outcomes are promising in view of real-life benefits. Future research should investigate the generalization of the observed benefits for signal detection to speech perception in everyday listening situations and determine the importance of coordination of bilateral speech processors and accentuation of envelope cues.

  5. Student diversity and implications for clinical competency development amongst domestic and international speech-language pathology students.

    PubMed

    Attrill, Stacie; Lincoln, Michelle; McAllister, Sue

    2012-06-01

    International students graduating from speech-language pathology university courses must achieve the same minimum competency standards as domestic students. This study aimed to collect descriptive information about the number, origin, and placement performance of international students as well as perceptions of the performance of international students on placement. University Clinical Education Coordinators (CECs), who manage clinical placements in eight undergraduate and six graduate entry programs across the 10 participating universities in Australia and New Zealand completed a survey about 3455 international and domestic speech-language pathology students. Survey responses were analysed quantitatively and qualitatively with non-parametric statistics and thematic analysis. Results indicated that international students came from a variety of countries, but with a regional focus on the countries of Central and Southern Asia. Although domestic students were noted to experience significantly less placement failure, fewer supplementary placements, and reduced additional placement support than international students, the effect size of these relationships was consistently small and therefore weak. CECs rated international students as more frequently experiencing difficulties with communication competencies on placement. However, CECs qualitative comments revealed that culturally and linguistically diverse (CALD) students may experience more difficulties with speech-language pathology competency development than international students. Students' CALD status should be included in future investigations of factors influencing speech-language pathology competency development.

  6. Telerehabilitation, virtual therapists, and acquired neurologic speech and language disorders.

    PubMed

    Cherney, Leora R; van Vuuren, Sarel

    2012-08-01

    Telerehabilitation (telerehab) offers cost-effective services that potentially can improve access to care for those with acquired neurologic communication disorders. However, regulatory issues including licensure, reimbursement, and threats to privacy and confidentiality hinder the routine implementation of telerehab services into the clinical setting. Despite these barriers, rapid technological advances and a growing body of research regarding the use of telerehab applications support its use. This article reviews the evidence related to acquired neurologic speech and language disorders in adults, focusing on studies that have been published since 2000. Research studies have used telerehab systems to assess and treat disorders including dysarthria, apraxia of speech, aphasia, and mild Alzheimer disease. They show that telerehab is a valid and reliable vehicle for delivering speech and language services. The studies represent a progression of technological advances in computing, Internet, and mobile technologies. They range on a continuum from working synchronously (in real-time) with a speech-language pathologist to working asynchronously (offline) with a stand-in virtual therapist. One such system that uses a virtual therapist for the treatment of aphasia, the Web-ORLA™ (Rehabilitation Institute of Chicago, Chicago, IL) system, is described in detail. Future directions for the advancement of telerehab for clinical practice are discussed. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  7. Availability of binaural cues for pediatric bilateral cochlear implant recipients.

    PubMed

    Sheffield, Sterling W; Haynes, David S; Wanna, George B; Labadie, Robert F; Gifford, René H

    2015-03-01

    Bilateral implant recipients theoretically have access to binaural cues. Research in postlingually deafened adults with cochlear implants (CIs) indicates minimal evidence for true binaural hearing. Congenitally deafened children who experience spatial hearing with bilateral CIs, however, might perceive binaural cues in the CI signal differently. There is limited research examining binaural hearing in children with CIs, and the few published studies are limited by the use of unrealistic speech stimuli and background noise. The purposes of this study were to (1) replicate our previous study of binaural hearing in postlingually deafened adults with AzBio sentences in prelingually deafened children with the pediatric version of the AzBio sentences, and (2) replicate previous studies of binaural hearing in children with CIs using more open-set sentences and more realistic background noise (i.e., multitalker babble). The study was a within-participant, repeated-measures design. The study sample consisted of 14 children with bilateral CIs with at least 25 mo of listening experience. Speech recognition was assessed using sentences presented in multitalker babble at a fixed signal-to-noise ratio. Test conditions included speech at 0° with noise presented at 0° (S0N0), on the side of the first CI (90° or 270°) (S0N1stCI), and on the side of the second CI (S0N2ndCI) as well as speech presented at 0° with noise presented semidiffusely from eight speakers at 45° intervals. Estimates of summation, head shadow, squelch, and spatial release from masking were calculated. Results of test conditions commonly reported in the literature (S0N0, S0N1stCI, S0N2ndCI) are consistent with results from previous research in adults and children with bilateral CIs, showing minimal summation and squelch but typical head shadow and spatial release from masking. However, bilateral benefit over the better CI with speech at 0° was much larger with semidiffuse noise. Congenitally deafened children with CIs have similar availability of binaural hearing cues to postlingually deafened adults with CIs within the same experimental design. It is possible that the use of realistic listening environments, such as semidiffuse background noise as in Experiment II, would reveal greater binaural hearing benefit for bilateral CI recipients. Future research is needed to determine whether (1) availability of binaural cues for children correlates with interaural time and level differences, (2) different listening environments are more sensitive to binaural hearing benefits, and (3) differences exist between pediatric bilateral recipients receiving implants in the same or sequential surgeries. American Academy of Audiology.

  8. Children with Specific Language Impairment and Their Families: A Future View of Nature Plus Nurture and New Technologies for Comprehensive Language Intervention Strategies.

    PubMed

    Rice, Mabel L

    2016-11-01

    Future perspectives on children with language impairments are framed from what is known about children with specific language impairment (SLI). A summary of the current state of services is followed by discussion of how these children can be overlooked and misunderstood and consideration of why it is so hard for some children to acquire language when it is effortless for most children. Genetic influences are highlighted, with the suggestion that nature plus nurture should be considered in present as well as future intervention approaches. A nurture perspective highlights the family context of the likelihood of SLI for some of the children. Future models of the causal pathways may provide more specific information to guide gene-treatment decisions, in ways parallel to current personalized medicine approaches. Future treatment options can build on the potential of electronic technologies and social media to provide personalized treatment methods available at a time and place convenient for the person to use as often as desired. The speech-language pathologist could oversee a wide range of treatment options and monitor evidence provided electronically to evaluate progress and plan future treatment steps. Most importantly, future methods can provide lifelong language acquisition activities that maintain the privacy and dignity of persons with language impairment, and in so doing will in turn enhance the effectiveness of speech-language pathologists. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  9. A Speech Recognition-based Solution for the Automatic Detection of Mild Cognitive Impairment from Spontaneous Speech

    PubMed Central

    Tóth, László; Hoffmann, Ildikó; Gosztolya, Gábor; Vincze, Veronika; Szatlóczki, Gréta; Bánréti, Zoltán; Pákáski, Magdolna; Kálmán, János

    2018-01-01

    Background: Even today the reliable diagnosis of the prodromal stages of Alzheimer’s disease (AD) remains a great challenge. Our research focuses on the earliest detectable indicators of cognitive de-cline in mild cognitive impairment (MCI). Since the presence of language impairment has been reported even in the mild stage of AD, the aim of this study is to develop a sensitive neuropsychological screening method which is based on the analysis of spontaneous speech production during performing a memory task. In the future, this can form the basis of an Internet-based interactive screening software for the recognition of MCI. Methods: Participants were 38 healthy controls and 48 clinically diagnosed MCI patients. The provoked spontaneous speech by asking the patients to recall the content of 2 short black and white films (one direct, one delayed), and by answering one question. Acoustic parameters (hesitation ratio, speech tempo, length and number of silent and filled pauses, length of utterance) were extracted from the recorded speech sig-nals, first manually (using the Praat software), and then automatically, with an automatic speech recogni-tion (ASR) based tool. First, the extracted parameters were statistically analyzed. Then we applied machine learning algorithms to see whether the MCI and the control group can be discriminated automatically based on the acoustic features. Results: The statistical analysis showed significant differences for most of the acoustic parameters (speech tempo, articulation rate, silent pause, hesitation ratio, length of utterance, pause-per-utterance ratio). The most significant differences between the two groups were found in the speech tempo in the delayed recall task, and in the number of pauses for the question-answering task. The fully automated version of the analysis process – that is, using the ASR-based features in combination with machine learning - was able to separate the two classes with an F1-score of 78.8%. Conclusion: The temporal analysis of spontaneous speech can be exploited in implementing a new, auto-matic detection-based tool for screening MCI for the community. PMID:29165085

  10. A Speech Recognition-based Solution for the Automatic Detection of Mild Cognitive Impairment from Spontaneous Speech.

    PubMed

    Toth, Laszlo; Hoffmann, Ildiko; Gosztolya, Gabor; Vincze, Veronika; Szatloczki, Greta; Banreti, Zoltan; Pakaski, Magdolna; Kalman, Janos

    2018-01-01

    Even today the reliable diagnosis of the prodromal stages of Alzheimer's disease (AD) remains a great challenge. Our research focuses on the earliest detectable indicators of cognitive decline in mild cognitive impairment (MCI). Since the presence of language impairment has been reported even in the mild stage of AD, the aim of this study is to develop a sensitive neuropsychological screening method which is based on the analysis of spontaneous speech production during performing a memory task. In the future, this can form the basis of an Internet-based interactive screening software for the recognition of MCI. Participants were 38 healthy controls and 48 clinically diagnosed MCI patients. The provoked spontaneous speech by asking the patients to recall the content of 2 short black and white films (one direct, one delayed), and by answering one question. Acoustic parameters (hesitation ratio, speech tempo, length and number of silent and filled pauses, length of utterance) were extracted from the recorded speech signals, first manually (using the Praat software), and then automatically, with an automatic speech recognition (ASR) based tool. First, the extracted parameters were statistically analyzed. Then we applied machine learning algorithms to see whether the MCI and the control group can be discriminated automatically based on the acoustic features. The statistical analysis showed significant differences for most of the acoustic parameters (speech tempo, articulation rate, silent pause, hesitation ratio, length of utterance, pause-per-utterance ratio). The most significant differences between the two groups were found in the speech tempo in the delayed recall task, and in the number of pauses for the question-answering task. The fully automated version of the analysis process - that is, using the ASR-based features in combination with machine learning - was able to separate the two classes with an F1-score of 78.8%. The temporal analysis of spontaneous speech can be exploited in implementing a new, automatic detection-based tool for screening MCI for the community. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  11. Transcranial direct current stimulation over left inferior frontal cortex improves speech fluency in adults who stutter.

    PubMed

    Chesters, Jennifer; Möttönen, Riikka; Watkins, Kate E

    2018-04-01

    See Crinion (doi:10.1093/brain/awy075) for a scientific commentary on this article.Stuttering is a neurodevelopmental condition affecting 5% of children, and persisting in 1% of adults. Promoting lasting fluency improvement in adults who stutter is a particular challenge. Novel interventions to improve outcomes are of value, therefore. Previous work in patients with acquired motor and language disorders reported enhanced benefits of behavioural therapies when paired with transcranial direct current stimulation. Here, we report the results of the first trial investigating whether transcranial direct current stimulation can improve speech fluency in adults who stutter. We predicted that applying anodal stimulation to the left inferior frontal cortex during speech production with temporary fluency inducers would result in longer-lasting fluency improvements. Thirty male adults who stutter completed a randomized, double-blind, controlled trial of anodal transcranial direct current stimulation over left inferior frontal cortex. Fifteen participants received 20 min of 1-mA stimulation on five consecutive days while speech fluency was temporarily induced using choral and metronome-timed speech. The other 15 participants received the same speech fluency intervention with sham stimulation. Speech fluency during reading and conversation was assessed at baseline, before and after the stimulation on each day of the 5-day intervention, and at 1 and 6 weeks after the end of the intervention. Anodal stimulation combined with speech fluency training significantly reduced the percentage of disfluent speech measured 1 week after the intervention compared with fluency intervention alone. At 6 weeks after the intervention, this improvement was maintained during reading but not during conversation. Outcome scores at both post-intervention time points on a clinical assessment tool (the Stuttering Severity Instrument, version 4) also showed significant improvement in the group receiving transcranial direct current stimulation compared with the sham group, in whom fluency was unchanged from baseline. We conclude that transcranial direct current stimulation combined with behavioural fluency intervention can improve fluency in adults who stutter. Transcranial direct current stimulation thereby offers a potentially useful adjunct to future speech therapy interventions for this population, for whom fluency therapy outcomes are currently limited.

  12. Speech intelligibility and subjective benefit in single-sided deaf adults after cochlear implantation.

    PubMed

    Finke, Mareike; Strauß-Schier, Angelika; Kludt, Eugen; Büchner, Andreas; Illg, Angelika

    2017-05-01

    Treatment with cochlear implants (CIs) in single-sided deaf individuals started less than a decade ago. CIs can successfully reduce incapacitating tinnitus on the deaf ear and allow, so some extent, the restoration of binaural hearing. Until now, systematic evaluations of subjective CI benefit in post-lingually single-sided deaf individuals and analyses of speech intelligibility outcome for the CI in isolation have been lacking. For the prospective part of this study, the Bern Benefit in Single-Sided Deafness Questionnaire (BBSS) was administered to 48 single-sided deaf CI users to evaluate the subjectively perceived CI benefit across different listening situations. In the retrospective part, speech intelligibility outcome with the CI up to 12 month post-activation was compared between 100 single-sided deaf CI users and 125 bilaterally implanted CI users (2nd implant). The positive median ratings in the BBSS differed significantly from zero for all items suggesting that most individuals with single-sided deafness rate their CI as beneficial across listening situations. The speech perception scores in quiet and noise improved significantly over time in both groups of CI users. Speech intelligibility with the CI in isolation was significantly better in bilaterally implanted CI users (2nd implant) compared to the scores obtained from single-sided deaf CI users. Our results indicate that CI users with single-sided deafness can reach open set speech understanding with their CI in isolation, encouraging the extension of the CI indication to individuals with normal hearing on the contralateral ear. Compared to the performance reached with bilateral CI users' second implant, speech reception threshold are lower, indicating an aural preference and dominance of the normal hearing ear. The results from the BBSS propose good satisfaction with the CI across several listening situations. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Speech production in children with Down's syndrome: The effects of reading, naming and imitation.

    PubMed

    Knight, Rachael-Anne; Kurtz, Scilla; Georgiadou, Ioanna

    2015-01-01

    People with DS are known to have difficulties with expressive language, and often have difficulties with intelligibility. They often have stronger visual than verbal short-term memory skills and, therefore, reading has often been suggested as an intervention for speech and language in this population. However, there is as yet no firm evidence that reading can improve speech outcomes. This study aimed to compare reading, picture naming and repetition for the same 10 words, to identify if the speech of eight children with DS (aged 11-14 years) was more accurate, consistent and intelligible when reading. Results show that children were slightly, yet significantly, more accurate and intelligible when they read words compared with when they produced those words in naming or imitation conditions although the reduction in inconsistency was non-significant. The results of this small-scale study provide tentative support for previous claims about the benefits of reading for children with DS. The mechanisms behind a facilitatory effect of reading are considered, and directions are identified for future research.

  14. Partial maintenance of auditory-based cognitive training benefits in older adults

    PubMed Central

    Anderson, Samira; White-Schwoch, Travis; Choi, Hee Jae; Kraus, Nina

    2014-01-01

    The potential for short-term training to improve cognitive and sensory function in older adults has captured the public’s interest. Initial results have been promising. For example, eight weeks of auditory-based cognitive training decreases peak latencies and peak variability in neural responses to speech presented in a background of noise and instills gains in speed of processing, speech-in-noise recognition, and short-term memory in older adults. But while previous studies have demonstrated short-term plasticity in older adults, we must consider the long-term maintenance of training gains. To evaluate training maintenance, we invited participants from an earlier training study to return for follow-up testing six months after the completion of training. We found that improvements in response peak timing to speech in noise and speed of processing were maintained, but the participants did not maintain speech-in-noise recognition or memory gains. Future studies should consider factors that are important for training maintenance, including the nature of the training, compliance with the training schedule, and the need for booster sessions after the completion of primary training. PMID:25111032

  15. Audio-vocal responses of vocal fundamental frequency and formant during sustained vowel vocalizations in different noises.

    PubMed

    Lee, Shao-Hsuan; Hsiao, Tzu-Yu; Lee, Guo-She

    2015-06-01

    Sustained vocalizations of vowels [a], [i], and syllable [mə] were collected in twenty normal-hearing individuals. On vocalizations, five conditions of different audio-vocal feedback were introduced separately to the speakers including no masking, wearing supra-aural headphones only, speech-noise masking, high-pass noise masking, and broad-band-noise masking. Power spectral analysis of vocal fundamental frequency (F0) was used to evaluate the modulations of F0 and linear-predictive-coding was used to acquire first two formants. The results showed that while the formant frequencies were not significantly shifted, low-frequency modulations (<3 Hz) of F0 significantly increased with reduced audio-vocal feedback across speech sounds and were significantly correlated with auditory awareness of speakers' own voices. For sustained speech production, the motor speech controls on F0 may depend on a feedback mechanism while articulation should rely more on a feedforward mechanism. Power spectral analysis of F0 might be applied to evaluate audio-vocal control for various hearing and neurological disorders in the future. Copyright © 2015 Elsevier B.V. All rights reserved.

  16. A clinical assessment of cochlear implant recipient performance: implications for individualized map settings in specific environments.

    PubMed

    Hey, Matthias; Hocke, Thomas; Mauger, Stefan; Müller-Deile, Joachim

    2016-11-01

    Individual speech intelligibility was measured in quiet and noise for cochlear Implant recipients upgrading from the Freedom to the CP900 series sound processor. The postlingually deafened participants (n = 23) used either Nucleus CI24RE or CI512 cochlear implant, and currently wore a Freedom sound processor. A significant group mean improvement in speech intelligibility was found in quiet (Freiburg monosyllabic words at 50 dB SPL ) and in noise (adaptive Oldenburger sentences in noise) for the two CP900 series SmartSound programs compared to the Freedom program. Further analysis was carried out on individual's speech intelligibility outcomes in quiet and in noise. Results showed a significant improvement or decrement for some recipients when upgrading to the new programs. To further increase speech intelligibility outcomes when upgrading, an enhanced upgrade procedure is proposed that includes additional testing with different signal-processing schemes. Implications of this research are that future automated scene analysis and switching technologies could provide additional performance improvements by introducing individualized scene-dependent settings.

  17. Effects of a computer-based intervention program on the communicative functions of children with autism.

    PubMed

    Hetzroni, Orit E; Tannous, Juman

    2004-04-01

    This study investigated the use of computer-based intervention for enhancing communication functions of children with autism. The software program was developed based on daily life activities in the areas of play, food, and hygiene. The following variables were investigated: delayed echolalia, immediate echolalia, irrelevant speech, relevant speech, and communicative initiations. Multiple-baseline design across settings was used to examine the effects of the exposure of five children with autism to activities in a structured and controlled simulated environment on the communication manifested in their natural environment. Results indicated that after exposure to the simulations, all children produced fewer sentences with delayed and irrelevant speech. Most of the children engaged in fewer sentences involving immediate echolalia and increased the number of communication intentions and the amount of relevant speech they produced. Results indicated that after practicing in a controlled and structured setting that provided the children with opportunities to interact in play, food, and hygiene activities, the children were able to transfer their knowledge to the natural classroom environment. Implications and future research directions are discussed.

  18. Comprehension: an overlooked component in augmented language development.

    PubMed

    Sevcik, Rose A

    2006-02-15

    Despite the importance of children's receptive skills as a foundation for later productive word use, the role of receptive language traditionally has received very limited attention since the focus in linguistic development has centered on language production. For children with significant developmental disabilities and communication impairments, augmented language systems have been devised as a tool both for language input and output. The role of both speech and symbol comprehension skills is emphasized in this paper. Data collected from two longitudinal studies of children and youth with severe disabilities and limited speech serve as illustrations in this paper. The acquisition and use of the System for Augmenting Language (SAL) was studied in home and school settings. Communication behaviors of the children and youth and their communication partners were observed and language assessment measures were collected. Two patterns of symbol learning and achievement--beginning and advanced--were observed. Extant speech comprehension skills brought to the augmented language learning task impacted the participants' patterns of symbol learning and use. Though often overlooked, the importance of speech and symbol comprehension skills were underscored in the studies described. Future areas for research are identified.

  19. Factors affecting articulation skills in children with velocardiofacial syndrome and children with cleft palate or velopharyngeal dysfunction: A preliminary report

    PubMed Central

    Baylis, Adriane L.; Munson, Benjamin; Moller, Karlind T.

    2010-01-01

    Objective To examine the influence of speech perception, cognition, and implicit phonological learning on articulation skills of children with Velocardiofacial syndrome (VCFS) and children with cleft palate or velopharyngeal dysfunction (VPD). Design Cross-sectional group experimental design. Participants 8 children with VCFS and 5 children with non-syndromic cleft palate or VPD. Methods and Measures All children participated in a phonetic inventory task, speech perception task, implicit priming nonword repetition task, conversational sample, nonverbal intelligence test, and hearing screening. Speech tasks were scored for percentage of phonemes correctly produced. Group differences and relations among measures were examined using nonparametric statistics. Results Children in the VCFS group demonstrated significantly poorer articulation skills and lower standard scores of nonverbal intelligence compared to the children with cleft palate or VPD. There were no significant group differences in speech perception skills. For the implicit priming task, both groups of children were more accurate in producing primed nonwords than unprimed nonwords. Nonverbal intelligence and severity of velopharyngeal inadequacy for speech were correlated with articulation skills. Conclusions In this study, children with VCFS had poorer articulation skills compared to children with cleft palate or VPD. Articulation difficulties seen in the children with VCFS did not appear to be associated with speech perception skills or the ability to learn new phonological representations. Future research should continue to examine relationships between articulation, cognition, and velopharyngeal dysfunction in a larger sample of children with cleft palate and VCFS. PMID:18333642

  20. Neural representations and mechanisms for the performance of simple speech sequences

    PubMed Central

    Bohland, Jason W.; Bullock, Daniel; Guenther, Frank H.

    2010-01-01

    Speakers plan the phonological content of their utterances prior to their release as speech motor acts. Using a finite alphabet of learned phonemes and a relatively small number of syllable structures, speakers are able to rapidly plan and produce arbitrary syllable sequences that fall within the rules of their language. The class of computational models of sequence planning and performance termed competitive queuing (CQ) models have followed Lashley (1951) in assuming that inherently parallel neural representations underlie serial action, and this idea is increasingly supported by experimental evidence. In this paper we develop a neural model that extends the existing DIVA model of speech production in two complementary ways. The new model includes paired structure and content subsystems (cf. MacNeilage, 1998) that provide parallel representations of a forthcoming speech plan, as well as mechanisms for interfacing these phonological planning representations with learned sensorimotor programs to enable stepping through multi-syllabic speech plans. On the basis of previous reports, the model’s components are hypothesized to be localized to specific cortical and subcortical structures, including the left inferior frontal sulcus, the medial premotor cortex, the basal ganglia and thalamus. The new model, called GODIVA (Gradient Order DIVA), thus fills a void in current speech research by providing formal mechanistic hypotheses about both phonological and phonetic processes that are grounded by neuroanatomy and physiology. This framework also generates predictions that can be tested in future neuroimaging and clinical case studies. PMID:19583476

  1. Auditory perception bias in speech imitation

    PubMed Central

    Postma-Nilsenová, Marie; Postma, Eric

    2013-01-01

    In an experimental study, we explored the role of auditory perception bias in vocal pitch imitation. Psychoacoustic tasks involving a missing fundamental indicate that some listeners are attuned to the relationship between all the higher harmonics present in the signal, which supports their perception of the fundamental frequency (the primary acoustic correlate of pitch). Other listeners focus on the lowest harmonic constituents of the complex sound signal which may hamper the perception of the fundamental. These two listener types are referred to as fundamental and spectral listeners, respectively. We hypothesized that the individual differences in speakers' capacity to imitate F0 found in earlier studies, may at least partly be due to the capacity to extract information about F0 from the speech signal. Participants' auditory perception bias was determined with a standard missing fundamental perceptual test. Subsequently, speech data were collected in a shadowing task with two conditions, one with a full speech signal and one with high-pass filtered speech above 300 Hz. The results showed that perception bias toward fundamental frequency was related to the degree of F0 imitation. The effect was stronger in the condition with high-pass filtered speech. The experimental outcomes suggest advantages for fundamental listeners in communicative situations where F0 imitation is used as a behavioral cue. Future research needs to determine to what extent auditory perception bias may be related to other individual properties known to improve imitation, such as phonetic talent. PMID:24204361

  2. The nature of articulation errors in Egyptian Arabic-speaking children with velopharyngeal insufficiency due to cleft palate.

    PubMed

    Abou-Elsaad, Tamer; Baz, Hemmat; Afsah, Omayma; Mansy, Alzahraa

    2015-09-01

    Even with early surgical repair, the majority of cleft palate children demonstrate articulation errors and have typical cleft palate speech. Was to determine the nature of articulation errors of Arabic consonants in Egyptian Arabic-speaking children with velopharyngeal insufficiency (VPI). Thirty Egyptian Arabic-speaking children with VPI due to cleft palate (whether primary repaired or secondary repaired) were studied. Auditory perceptual assessment (APA) of children speech was conducted. Nasopharyngoscopy was done to assess the velopharyngeal port (VPP) movements while the child was repeating speech tasks. Mansoura Arabic Articulation test (MAAT) was performed to analyze the consonants articulation of these children. The most frequent type of articulatory errors observed was substitution, more specifically, backing. Pharyngealization of anterior fricatives was the most frequent substitution, especially for the /s/ sound. The most frequent substituting sounds for other sounds were /ʔ/ followed by /k/ and /n/ sounds. Significant correlations were found between the degrees of the open nasality and VPP closure and the articulation errors. On the other hand, the sounds (/ʔ/,/ħ/,/ʕ/,/n/,/w/,/j/) were normally articulated in all studied group. The determination of articulation errors in VPI children could guide the therapists for designing appropriate speech therapy programs for these cases. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  3. Language Disorders in Adolescents: Current Needs and Future Directions.

    PubMed

    Nippold, Marilyn A

    2016-11-01

    Adolescents with developmental language disorders often do not receive the type of intervention that would improve their ability to speak, listen, read, and write effectively. Part of the problem is that many of these young people show no obvious symptoms of a language disorder, yet they struggle on a daily basis to succeed at school-related tasks that require a sophisticated level of language development. This article discusses some of the challenges these students face and makes suggestions for what could be done to address the issues. These suggestions include continuing the effort to advocate strongly for the rights of adolescents, increasing collaboration between speech-language pathologists and other professionals in the schools, and making changes to training programs in communication sciences and disorders to better prepare future speech-language pathologists to work with adolescents. Thieme Medical Publishers 333 Seventh Avenue, New York, NY 10001, USA.

  4. Family-centered services for children with complex communication needs: the practices and beliefs of school-based speech-language pathologists.

    PubMed

    Mandak, Kelsey; Light, Janice

    2018-06-01

    This study used an online focus group to examine the beliefs and practices of school-based speech-language pathologists (SLPs) who served children with complex communication needs regarding their provision of family-centered services. Participants revealed that despite their desire for family involvement and reported beliefs in the importance of family-centered services, there were barriers in place that often limited family-centered service provision. Across the SLPs, many were dissatisfied with their current provision of family-centered services. The SLPs varied in their reported practices, with some reporting family-centered services and others, professional-centered services. Future research is recommended in order to investigate which factors contribute to the variation among SLPs and how the variation impacts children who require augmentative and alternative communication (AAC) and their families. Potential clinical implications for in-service and pre-service SLPs are discussed to improve future family-centered AAC services.

  5. Importance of age and postimplantation experience on speech perception measures in children with sequential bilateral cochlear implants.

    PubMed

    Peters, B Robert; Litovsky, Ruth; Parkinson, Aaron; Lake, Jennifer

    2007-08-01

    Clinical trials in which children received bilateral cochlear implants in sequential operations were conducted to analyze the extent to which bilateral implantation offers benefits on a number of measures. The present investigation was particularly focused on measuring the effects of age at implantation and experience after activation of the second implant on speech perception performance. Thirty children aged 3 to 13 years were recipients of 2 cochlear implants, received in sequential operations, a minimum of 6 months apart. All children received their first implant before 5 years of age and had acquired speech perception capabilities with the first device. They were divided into 3 age groups on the basis of age at time of second ear implantation: Group I, 3 to 5 years; Group II, 5.1 to 8 years; and Group III, 8.1 to 13 years. Speech perception measures in quiet included the Multisyllabic Lexical Neighborhood Test (MLNT) for Group I, the Lexical Neighborhood Test (LNT) for Groups II and III, and the Hearing In Noise Test for Children (HINT-C) sentences in quiet for Group III. Speech perception in noise was assessed using the Children's Realistic Intelligibility and Speech Perception (CRISP) test. Testing was performed preoperatively and again postactivation of the second implant at 3, 6, and 12 months (CRISP at 3 and 9 mo) in both the unilateral and bilateral conditions in a repeated-measures study design. Two-way repeated-measures analysis of variance was used to analyze statistical significance among device configurations and performance over time. US Multicenter. Results for speech perception in quiet show that children implanted sequentially acquire open-set speech perception in the second ear relatively quickly (within 6 mo). However, children younger than 8 years do so more rapidly and to a higher level of speech perception ability at 12 months than older children (mean second ear MLNT/LNT scores at 12 months: Group I, 83.9%; range, 71-96%; Group II, 59.5%; range, 40-88%; Group III, 32%; range, 12-56%). The second-ear mean HINT-C score for Group III children remained far less than that of the first ear even after 12 months of device use (44 versus 89%; t, 6.48; p<0.001; critical value, 0.025). Speech intelligibility for spondees in noise was significantly better under bilateral conditions than with either ear alone when all children were analyzed as a single group and for Group III children. At the 9-month test interval, performance in the bilateral configuration was significantly better for all noise conditions (13.2% better for noise at first cochlear implant, 6.8% better for the noise front and noise at second cochlear implant conditions, t=2.32, p=0.024, critical level=0.05 for noise front; t=3.75, p<0.0001, critical level=0.05 for noise at first implant; t=2.73, p = 0.008, critical level=0.05 for noise at second implant side). The bilateral benefit in noise increased with time from 3 to 9 months after activation of the second implant. This bilateral advantage is greatest when noise is directed toward the first implanted ear, indicating that the head shadow effect is the most effective binaural mechanism. The bilateral condition produced small improvements in speech perception in quiet and for individual Group I and Group II patient results in noise that, in view of the relatively small number of subjects tested, do not reach statistical significance. Sequential bilateral cochlear implantation in children of diverse ages has the potential to improve speech perception abilities in the second implanted ear and to provide access to the use of binaural mechanisms such as the head shadow effect. The improvement unfolds over time and continues to grow during the 6 to 12 months after activation of the second implant. Younger children in this study achieved higher open-set speech perception scores in the second ear, but older children still demonstrate bilateral benefit in noise. Determining the long-term impact and cost-effectiveness that results from such potential capabilities in bilaterally implanted children requires additional study with larger groups of subjects and more prolonged monitoring.

  6. Speech sound articulation abilities of preschool-age children who stutter.

    PubMed

    Clark, Chagit E; Conture, Edward G; Walden, Tedra A; Lambert, Warren E

    2013-12-01

    The purpose of this study was to assess the association between speech sound articulation and childhood stuttering in a relatively large sample of preschool-age children who do and do not stutter, using the Goldman-Fristoe Test of Articulation-2 (GFTA-2; Goldman & Fristoe, 2000). Participants included 277 preschool-age children who do (CWS; n=128, 101 males) and do not stutter (CWNS; n=149, 76 males). Generalized estimating equations (GEE) were performed to assess between-group (CWS versus CWNS) differences on the GFTA-2. Additionally, within-group correlations were performed to explore the relation between CWS' speech sound articulation abilities and their stuttering frequency and severity, as well as their sound prolongation index (SPI; Schwartz & Conture, 1988). No significant differences were found between the articulation scores of preschool-age CWS and CWNS. However, there was a small gender effect for the 5-year-old age group, with girls generally exhibiting better articulation scores than boys. Additional findings indicated no relation between CWS' speech sound articulation abilities and their stuttering frequency, severity, or SPI. Findings suggest no apparent association between speech sound articulation-as measured by one standardized assessment (GFTA-2)-and childhood stuttering for this sample of preschool-age children (N=277). After reading this article, the reader will be able to: (1) discuss salient issues in the articulation literature relative to children who stutter; (2) compare/contrast the present study's methodologies and main findings to those of previous studies that investigated the association between childhood stuttering and speech sound articulation; (3) identify future research needs relative to the association between childhood stuttering and speech sound development; (4) replicate the present study's methodology to expand this body of knowledge. Copyright © 2013 Elsevier Inc. All rights reserved.

  7. The affective reactivity of psychotic speech: The role of internal source monitoring in explaining increased thought disorder under emotional challenge.

    PubMed

    de Sousa, Paulo; Sellwood, William; Spray, Amy; Bentall, Richard P

    2016-04-01

    Thought disorder (TD) has been shown to vary in relation to negative affect. Here we examine the role internal source monitoring (iSM, i.e. ability to discriminate between inner speech and verbalized speech) in TD and whether changes in iSM performance are implicated in the affective reactivity effect (deterioration of TD when participants are asked to talk about emotionally-laden topics). Eighty patients diagnosed with schizophrenia-spectrum disorder and thirty healthy controls received interviews that promoted personal disclosure (emotionally salient) and interviews on everyday topics (non-salient) on separate days. During the interviews, participants were tested on iSM, self-reported affect and immediate auditory recall. Patients had more TD, poorer ability to discriminate between inner and verbalized speech, poorer immediate auditory recall and reported more negative affect than controls. Both groups displayed more TD and negative affect in salient interviews but only patients showed poorer performance on iSM. Immediate auditory recall did not change significantly across affective conditions. In patients, the relationship between self-reported negative affect and TD was mediated by deterioration in the ability to discriminate between inner speech and speech that was directed to others and socially shared (performance on the iSM) in both interviews. Furthermore, deterioration in patients' performance on iSM across conditions significantly predicted deterioration in TD across the interviews (affective reactivity of speech). Poor iSM is significantly associated with TD. Negative affect, leading to further impaired iSM, leads to increased TD in patients with psychosis. Avenues for future research as well as clinical implications of these findings are discussed. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.

  8. Determining the importance of fundamental hearing aid attributes.

    PubMed

    Meister, Hartmut; Lausberg, Isabel; Kiessling, Juergen; Walger, Martin; von Wedel, Hasso

    2002-07-01

    To determine the importance of fundamental hearing aid attributes and to elicit measures of satisfaction and dissatisfaction. A prospective study based on a survey using a decompositional approach of preference measurement (conjoint analysis). Ear, nose, and throat university hospitals in Cologne and Giessen; various branches of hearing aid dispensers. A random sample of 175 experienced hearing aid users aged 20 to 91 years (mean age, 61 yr) recruited at two different sites. Relative importance of different hearing aid attributes, satisfaction and dissatisfaction with hearing aid attributes. Of the six fundamental hearing aid attributes assessed by the hearing aid users, the two features concerning speech perception attained the highest relative importance (25% speech in quiet, 27% speech in noise). The remaining four attributes (sound quality, handling, feedback, localization) had significantly lower values in a narrow range of 10 to 12%. Comparison of different subgroups of hearing aid wearers based on sociodemographic and user-specific data revealed a large interindividual scatter of the preferences for the attributes. A similar examination with 25 clinicians revealed overestimation of the importance of the attributes commonly associated with problems. Moreover, examination of satisfaction showed that speech in noise was the most frequent source of dissatisfaction (30% of all statements), whereas the subjects were satisfied with speech in quiet. The results emphasize the high importance of attributes related to speech perception. Speech discrimination in noise was the most important but also the most frequent source of negative statements. This attribute will be the outstanding parameter of future developments. Appropriate handling becomes an important factor for elderly subjects. However, because of the large interindividual scatter of data, the preferences of different hearing aid users were hardly predictable, giving evidence of multifactorial influences.

  9. An Innovative Speech-Based User Interface for Smarthomes and IoT Solutions to Help People with Speech and Motor Disabilities.

    PubMed

    Malavasi, Massimiliano; Turri, Enrico; Atria, Jose Joaquin; Christensen, Heidi; Marxer, Ricard; Desideri, Lorenzo; Coy, Andre; Tamburini, Fabio; Green, Phil

    2017-01-01

    A better use of the increasing functional capabilities of home automation systems and Internet of Things (IoT) devices to support the needs of users with disability, is the subject of a research project currently conducted by Area Ausili (Assistive Technology Area), a department of Polo Tecnologico Regionale Corte Roncati of the Local Health Trust of Bologna (Italy), in collaboration with AIAS Ausilioteca Assistive Technology (AT) Team. The main aim of the project is to develop experimental low cost systems for environmental control through simplified and accessible user interfaces. Many of the activities are focused on automatic speech recognition and are developed in the framework of the CloudCAST project. In this paper we report on the first technical achievements of the project and discuss future possible developments and applications within and outside CloudCAST.

  10. Perception of suprasegmental features of speech by children with cochlear implants and children with hearing AIDS.

    PubMed

    Most, Tova; Peled, Miriam

    2007-01-01

    This study assessed perception of suprasegmental features of speech by 30 prelingual children with sensorineural hearing loss. Ten children had cochlear implants (CIs), and 20 children wore hearing aids (HA): 10 with severe hearing loss and 10 with profound hearing loss. Perception of intonation, syllable stress, word emphasis, and word pattern was assessed. Results revealed that the two HA groups significantly outperformed the CI group in perceiving both intonation and stress. Within each group, word pattern was perceived best, and then intonation and emphasis, with syllable stress perceived poorest. No significant correlation emerged between age at implantation and perception of the various suprasegmental features, possibly due to participants' relatively late age at implantation. Results indicated that CI use did not show an advantage over HA use in the perception of suprasegmental features of speech. Future research should continue to explore variables that might improve this perception.

  11. Challenges in discriminating profanity from hate speech

    NASA Astrophysics Data System (ADS)

    Malmasi, Shervin; Zampieri, Marcos

    2018-03-01

    In this study, we approach the problem of distinguishing general profanity from hate speech in social media, something which has not been widely considered. Using a new dataset annotated specifically for this task, we employ supervised classification along with a set of features that includes ?-grams, skip-grams and clustering-based word representations. We apply approaches based on single classifiers as well as more advanced ensemble classifiers and stacked generalisation, achieving the best result of ? accuracy for this 3-class classification task. Analysis of the results reveals that discriminating hate speech and profanity is not a simple task, which may require features that capture a deeper understanding of the text not always possible with surface ?-grams. The variability of gold labels in the annotated data, due to differences in the subjective adjudications of the annotators, is also an issue. Other directions for future work are discussed.

  12. Gender typicality in children's speech: A comparison of boys with and without gender identity disorder.

    PubMed

    Munson, Benjamin; Crocker, Laura; Pierrehumbert, Janet B; Owen-Anderson, Allison; Zucker, Kenneth J

    2015-04-01

    This study examined whether boys with gender identity disorder (GID) produced less prototypically male speech than control boys without GID, a possibility that has been suggested by clinical observations. Two groups of listeners participated in tasks where they rated the gender typicality of single words (group 1) or sentences (group 2) produced by 15 5-13 year old boys with GID and 15 age-matched boys without GID. Detailed acoustic analyses of the stimuli were also conducted. Boys with GID were rated as less boy-like than boys without GID. In the experiment using sentence stimuli, these group differences were larger than in the experiment using single-word stimuli. Listeners' ratings were predicted by a variety of acoustic parameters, including ones that differ between the two groups and ones that are stereotypically associated with adult men's and women's speech. Future research should examine how these variants are acquired.

  13. Advances in EPG for treatment and research: an illustrative case study.

    PubMed

    Scobbie, James M; Wood, Sara E; Wrench, Alan A

    2004-01-01

    Electropalatography (EPG), a technique which reveals tongue-palate contact patterns over time, is a highly effective tool for speech research. We report here on recent developments by Articulate Instruments Ltd. These include hardware for Windows-based computers, backwardly compatible (with Reading EPG3) software systems for clinical intervention and laboratory-based analysis for EPG and acoustic data, and an enhanced clinical interface with client and file management tools. We focus here on a single case study of a child aged 10+/-years who had been diagnosed with an intractable speech disorder possibly resulting ultimately from a complete cleft of hard and soft palate. We illustrate how assessment, diagnosis and treatment of the intractable speech disorder are undertaken using this new generation of instrumental phonetic support. We also look forward to future developments in articulatory phonetics that will link EPG with ultrasound for research and clinical communities.

  14. Influence of Syllable Structure on L2 Auditory Word Learning

    ERIC Educational Resources Information Center

    Hamada, Megumi; Goya, Hideki

    2015-01-01

    This study investigated the role of syllable structure in L2 auditory word learning. Based on research on cross-linguistic variation of speech perception and lexical memory, it was hypothesized that Japanese L1 learners of English would learn English words with an open-syllable structure without consonant clusters better than words with a…

  15. The International Research Forum in Information Science Proceedings (4th, Boras, Sweden, September 14-16, 1981).

    ERIC Educational Resources Information Center

    Friberg, Ingegerd, Ed.

    The Fourth International Research Forum in Information Science (IRFIS 4) focused on two themes: user-oriented research in information science and experimental and practical investigations of the mediation and use of information. An opening speech by Bjorn Tell reviewed information research policy in the United States, France, England, Japan, the…

  16. The National Conference on Professional and Personal Renewal for Faculty (2nd, Atlanta, Georgia, April 13-15, 1989).

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Office of Instructional Development.

    This document reports on the second of a projected series of conferences, for University of Georgia faculty, having the theme of "faculty renewal" (rather than the older "faculty development") as the major focus. Following a brief introduction, the proceedings open with the keynote speech, "Meeting the Challenge of…

  17. Regional Development and Vocational Training. The Problems of Evaluation. Colloquium (Limoges, France, June 18-19, 1990).

    ERIC Educational Resources Information Center

    Pierret, M.

    1990-01-01

    This paper summarizes the proceedings of a conference held to discuss evaluation in vocational education. The agenda included welcome speeches and the introductory address by Robert Savy, President of the Regional Council of Limousin (France); an opening session with statements from representatives of the Commission of the European Communities…

  18. Rhetoric, Possibility, and Women's Status in Ancient Athens: Gorgias' and Isocrates' Encomiums of Helen.

    ERIC Educational Resources Information Center

    Biesecker, Susan L.

    1992-01-01

    Critiques teleological narrative structures implicit in the work of some classical historians, adopting instead a multilayered historiographical method. Argues that a law instituted in 451/450 B.C.E. by Pericles opened up a possibility for resisting women's exclusion from the public sphere. Reads Gorgias' and Isocrates' speeches on Helen of Troy…

  19. Opening the Doors of Communication: How to Choose Communication Tools for Your Child with Disabilities. Fourth Edition

    ERIC Educational Resources Information Center

    PACER Center, 2009

    2009-01-01

    Communication is important to all people. Through gestures, body language, writing, facial expressions, speech, and other means, people are able to share their thoughts and ideas, build relationships, and express their needs. When they cannot communicate, their behavior, learning, and sociability can all suffer. Fortunately, augmentative and…

  20. Speech and communication in Parkinson’s disease: a cross-sectional exploratory study in the UK

    PubMed Central

    Barnish, Maxwell S; Horton, Simon M C; Butterfint, Zoe R; Clark, Allan B; Atkinson, Rachel A; Deane, Katherine H O

    2017-01-01

    Objective To assess associations between cognitive status, intelligibility, acoustics and functional communication in PD. Design Cross-sectional exploratory study of functional communication, including a within-participants experimental design for listener assessment. Setting A major academic medical centre in the East of England, UK. Participants Questionnaire data were assessed for 45 people with Parkinson’s disease (PD), who had self-reported speech or communication difficulties and did not have clinical dementia. Acoustic and listener analyses were conducted on read and conversational speech for 20 people with PD and 20 familiar conversation partner controls without speech, language or cognitive difficulties. Main outcome measures Functional communication assessed by the Communicative Participation Item Bank (CPIB) and Communicative Effectiveness Survey (CES). Results People with PD had lower intelligibility than controls for both the read (mean difference 13.7%, p=0.009) and conversational (mean difference 16.2%, p=0.04) sentences. Intensity and pause were statistically significant predictors of intelligibility in read sentences. Listeners were less accurate identifying the intended emotion in the speech of people with PD (14.8% point difference across conditions, p=0.02) and this was associated with worse speaker cognitive status (16.7% point difference, p=0.04). Cognitive status was a significant predictor of functional communication using CPIB (F=8.99, p=0.005, η2 = 0.15) but not CES. Intelligibility in conversation sentences was a statistically significant predictor of CPIB (F=4.96, p=0.04, η2 = 0.19) and CES (F=13.65, p=0.002, η2 = 0.43). Read sentence intelligibility was not a significant predictor of either outcome. Conclusions Cognitive status was an important predictor of functional communication—the role of intelligibility was modest and limited to conversational and not read speech. Our results highlight the importance of focusing on functional communication as well as physical speech impairment in speech and language therapy (SLT) for PD. Our results could inform future trials of SLT techniques for PD. PMID:28554918

  1. Nonword repetition and nonword reading abilities in adults who do and do not stutter.

    PubMed

    Sasisekaran, Jayanthi

    2013-09-01

    In the present study a nonword repetition and a nonword reading task were used to investigate the behavioral (speech accuracy) and speech kinematic (movement variability measured as lip aperture variability index; speech duration) profiles of groups of young adults who do (AWS) and do not stutter (control). Participants were 9 AWS (8 males, Mean age=32.2, SD=14.7) and 9 age- and sex-matched control participants (Mean age=31.8, SD=14.6). For the nonword repetition task, participants were administered the Nonword Repetition Test (Dollaghan & Campbell, 1998). For the reading task, participants were required to read out target nonwords varying in length (6 vs. 11 syllables). Repeated measures analyses of variance were conducted to compare the groups in percent speech accuracy for both tasks; only for the nonword reading task, the groups were compared in movement variability and speech duration. The groups were comparable in percent accuracy in nonword repetition. Findings from nonword reading revealed a trend for the AWS to show a lower percent of accurate productions compared to the control group. AWS also showed significantly higher movement variability and longer speech durations compared to the control group in nonword reading. Some preliminary evidence for group differences in practice effect (seen as differences between the early vs. later 5 trials) was evident in speech duration. Findings suggest differences between AWS and control groups in phonemic encoding and/or speech motor planning and production. Findings from nonword repetition vs. reading highlight the need for careful consideration of nonword properties. At the end of this activity the reader will be able to: (a) summarize the literature on nonword repetition skills in adults who stutter, (b) describe processes underlying nonword repetition and nonword reading, (c) summarize whether or not adults who stutter differ from those who do not in the behavioral and kinematic markers of nonword reading performance, (d) discuss future directions for research. Copyright © 2013 Elsevier Inc. All rights reserved.

  2. Integration of auditory and somatosensory error signals in the neural control of speech movements.

    PubMed

    Feng, Yongqiang; Gracco, Vincent L; Max, Ludo

    2011-08-01

    We investigated auditory and somatosensory feedback contributions to the neural control of speech. In task I, sensorimotor adaptation was studied by perturbing one of these sensory modalities or both modalities simultaneously. The first formant (F1) frequency in the auditory feedback was shifted up by a real-time processor and/or the extent of jaw opening was increased or decreased with a force field applied by a robotic device. All eight subjects lowered F1 to compensate for the up-shifted F1 in the feedback signal regardless of whether or not the jaw was perturbed. Adaptive changes in subjects' acoustic output resulted from adjustments in articulatory movements of the jaw or tongue. Adaptation in jaw opening extent in response to the mechanical perturbation occurred only when no auditory feedback perturbation was applied or when the direction of adaptation to the force was compatible with the direction of adaptation to a simultaneous acoustic perturbation. In tasks II and III, subjects' auditory and somatosensory precision and accuracy were estimated. Correlation analyses showed that the relationships 1) between F1 adaptation extent and auditory acuity for F1 and 2) between jaw position adaptation extent and somatosensory acuity for jaw position were weak and statistically not significant. Taken together, the combined findings from this work suggest that, in speech production, sensorimotor adaptation updates the underlying control mechanisms in such a way that the planning of vowel-related articulatory movements takes into account a complex integration of error signals from previous trials but likely with a dominant role for the auditory modality.

  3. Working with culturally and linguistically diverse students and their families: perceptions and practices of school speech-language therapists in the United States.

    PubMed

    Maul, Christine A

    2015-01-01

    Speech and language therapists (SLTs) working in schools worldwide strive to deliver evidence-based services to diverse populations of students. Many suggestions have been made in the international professional literature regarding culturally competent delivery of speech and language services, but there has been limited qualitative investigation of practices school SLTs find to be most useful when modifying their approaches to meet the needs of culturally and linguistically diverse (CLD) students. To examine perceptions of nine school SLTs regarding modifications of usual practices when interacting with CLD students and their families; to compare reported practices with those suggested in professional literature; to draw clinical implications regarding the results; and to suggest future research to build a more extensive evidence base for culturally competent service delivery. For this qualitative research study, nine school SLTs in a diverse region of the USA were recruited to participate in a semi-structured interview designed to answer the question: What dominant themes, if any, can be found in SLTs' descriptions of how they modify their approaches, if at all, when interacting with CLD students and their family members? Analysis of data revealed the following themes: (1) language-a barrier and a bridge, (2) communicating through interpreters, (3) respect for cultural differences, and (4) positive experiences interacting with CLD family members. Participants reported making many modifications to their usual approaches that have been recommended as best practices in the international literature. However, some practices the SLTs reported to be effective were not emphasized or were not addressed at all in the literature. Practical implications of results are drawn and future research is suggested. © 2015 Royal College of Speech and Language Therapists.

  4. Proceedings 2011: Selected Papers from the Fifteenth College-Wide Conference for Students in Languages, Linguistics & Literature (15th, Honolulu, Hawai'i, April 16, 2011)

    ERIC Educational Resources Information Center

    Carey, Britany, Ed.; Sasayama, Shoko, Ed.

    2011-01-01

    The theme of this year's conference was "Check Your Vision for the Future" with papers highlighting emerging changes in language use and literature, as well as suggestions for improving language teaching and maintenance in the future. The plenary speech was given by Dr. Katie Drager, followed by student presentations. Following a preface…

  5. Making Microcomputers Accessible to Blind Persons.

    ERIC Educational Resources Information Center

    Ruconich, Sandra K.; And Others

    1986-01-01

    The article considers advantages and limitations of tactile, auditory, and visual means of microcomputer access for blind students. Discussed are electronic braille, paper braille, the Optacon, synthesized speech, and enlarged print. Improved multimedia access technology is predicted for the future. (Author/DB)

  6. A speech processing study using an acoustic model of a multiple-channel cochlear implant

    NASA Astrophysics Data System (ADS)

    Xu, Ying

    1998-10-01

    A cochlear implant is an electronic device designed to provide sound information for adults and children who have bilateral profound hearing loss. The task of representing speech signals as electrical stimuli is central to the design and performance of cochlear implants. Studies have shown that the current speech- processing strategies provide significant benefits to cochlear implant users. However, the evaluation and development of speech-processing strategies have been complicated by hardware limitations and large variability in user performance. To alleviate these problems, an acoustic model of a cochlear implant with the SPEAK strategy is implemented in this study, in which a set of acoustic stimuli whose psychophysical characteristics are as close as possible to those produced by a cochlear implant are presented on normal-hearing subjects. To test the effectiveness and feasibility of this acoustic model, a psychophysical experiment was conducted to match the performance of a normal-hearing listener using model- processed signals to that of a cochlear implant user. Good agreement was found between an implanted patient and an age-matched normal-hearing subject in a dynamic signal discrimination experiment, indicating that this acoustic model is a reasonably good approximation of a cochlear implant with the SPEAK strategy. The acoustic model was then used to examine the potential of the SPEAK strategy in terms of its temporal and frequency encoding of speech. It was hypothesized that better temporal and frequency encoding of speech can be accomplished by higher stimulation rates and a larger number of activated channels. Vowel and consonant recognition tests were conducted on normal-hearing subjects using speech tokens processed by the acoustic model, with different combinations of stimulation rate and number of activated channels. The results showed that vowel recognition was best at 600 pps and 8 activated channels, but further increases in stimulation rate and channel numbers were not beneficial. Manipulations of stimulation rate and number of activated channels did not appreciably affect consonant recognition. These results suggest that overall speech performance may improve by appropriately increasing stimulation rate and number of activated channels. Future revision of this acoustic model is necessary to provide more accurate amplitude representation of speech.

  7. MCA-NMF: Multimodal Concept Acquisition with Non-Negative Matrix Factorization

    PubMed Central

    Mangin, Olivier; Filliat, David; ten Bosch, Louis; Oudeyer, Pierre-Yves

    2015-01-01

    In this paper we introduce MCA-NMF, a computational model of the acquisition of multimodal concepts by an agent grounded in its environment. More precisely our model finds patterns in multimodal sensor input that characterize associations across modalities (speech utterances, images and motion). We propose this computational model as an answer to the question of how some class of concepts can be learnt. In addition, the model provides a way of defining such a class of plausibly learnable concepts. We detail why the multimodal nature of perception is essential to reduce the ambiguity of learnt concepts as well as to communicate about them through speech. We then present a set of experiments that demonstrate the learning of such concepts from real non-symbolic data consisting of speech sounds, images, and motions. Finally we consider structure in perceptual signals and demonstrate that a detailed knowledge of this structure, named compositional understanding can emerge from, instead of being a prerequisite of, global understanding. An open-source implementation of the MCA-NMF learner as well as scripts and associated experimental data to reproduce the experiments are publicly available. PMID:26489021

  8. Influence of signal processing strategy in auditory abilities.

    PubMed

    Melo, Tatiana Mendes de; Bevilacqua, Maria Cecília; Costa, Orozimbo Alves; Moret, Adriane Lima Mortari

    2013-01-01

    The signal processing strategy is a parameter that may influence the auditory performance of cochlear implant and is important to optimize this parameter to provide better speech perception, especially in difficult listening situations. To evaluate the individual's auditory performance using two different signal processing strategy. Prospective study with 11 prelingually deafened children with open-set speech recognition. A within-subjects design was used to compare performance with standard HiRes and HiRes 120 in three different moments. During test sessions, subject's performance was evaluated by warble-tone sound-field thresholds, speech perception evaluation, in quiet and in noise. In the silence, children S1, S4, S5, S7 showed better performance with the HiRes 120 strategy and children S2, S9, S11 showed better performance with the HiRes strategy. In the noise was also observed that some children performed better using the HiRes 120 strategy and other with HiRes. Not all children presented the same pattern of response to the different strategies used in this study, which reinforces the need to look at optimizing cochlear implant clinical programming.

  9. Minimally invasive surgical method to detect sound processing in the cochlear apex by optical coherence tomography

    NASA Astrophysics Data System (ADS)

    Ramamoorthy, Sripriya; Zhang, Yuan; Petrie, Tracy; Fridberger, Anders; Ren, Tianying; Wang, Ruikang; Jacques, Steven L.; Nuttall, Alfred L.

    2016-02-01

    Sound processing in the inner ear involves separation of the constituent frequencies along the length of the cochlea. Frequencies relevant to human speech (100 to 500 Hz) are processed in the apex region. Among mammals, the guinea pig cochlear apex processes similar frequencies and is thus relevant for the study of speech processing in the cochlea. However, the requirement for extensive surgery has challenged the optical accessibility of this area to investigate cochlear processing of signals without significant intrusion. A simple method is developed to provide optical access to the guinea pig cochlear apex in two directions with minimal surgery. Furthermore, all prior vibration measurements in the guinea pig apex involved opening an observation hole in the otic capsule, which has been questioned on the basis of the resulting changes to cochlear hydrodynamics. Here, this limitation is overcome by measuring the vibrations through the unopened otic capsule using phase-sensitive Fourier domain optical coherence tomography. The optically and surgically advanced method described here lays the foundation to perform minimally invasive investigation of speech-related signal processing in the cochlea.

  10. SUS users' perception: a speech-language pathology approach based on health promotion.

    PubMed

    Cunha, Jenane Topanotti da; Massi, Giselle; Guarinello, Ana Cristina; Pereira, Francine Martins

    2016-01-01

    This study aimed to analyze the perceptions of users of the Brazilian Unified Health System (SUS) about the treatment Center where they were assisted, as well as about the speech-language pathology services rendered by this Center. This is a transversal study composed of an interview with 26 open questions and 14 closed questions applied to 111 individuals who were assisted at the SUS Center in August 2013. The quantitative content analysis was conducted through the use of the GraphPadPrisma 5.1, Statistic Package for Social Sciences (SPSS) 15.0 software and the application of the D'agostino & Person, F-test and chi-squared test. Most participants reported a positive perception about the facilities and staff of the treatment center. They were also positive about the waiting time and the speech-language pathologists' explanations and conduct, especially in the audiology department. Most responses from participants were short and did not present an argumentative context. The treatment center received a high approval rating by most users. The audiology department had better grades than the clinical services related to language and oral motor pathologies.

  11. Foucault's "fearless speech" and the transformation and mentoring of medical students.

    PubMed

    Papadimos, Thomas J; Murray, Stuart J

    2008-04-17

    In his six 1983 lectures published under the title, Fearless Speech (2001), Michel Foucault developed the theme of free speech and its relation to frankness, truth-telling, criticism, and duty. Derived from the ancient Greek word parrhesia, Foucault's analysis of free speech is relevant to the mentoring of medical students. This is especially true given the educational and social need to transform future physicians into able citizens who practice a fearless freedom of expression on behalf of their patients, the public, the medical profession, and themselves in the public and political arena. In this paper, we argue that Foucault's understanding of free speech, or parrhesia, should be read as an ethical response to the American Medical Association's recent educational effort, Initiative to Transform Medical Education (ITME): Recommendations for change in the system of medical education (2007). In this document, the American Medical Association identifies gaps in medical education, emphasizing the need to enhance health system safety and quality, to improve education in training institutions, and to address the inadequacy of physician preparedness in new content areas. These gaps, and their relationship to the ITME goal of promoting excellence in patient care by implementing reform in the US system of medical education, call for a serious consideration and use of Foucault's parrhesia in the way that medical students are trained and mentored.

  12. Technology and the evolution of clinical methods for stuttering.

    PubMed

    Packman, Ann; Meredith, Grant

    2011-06-01

    The World Wide Web (WWW) was 20 years old last year. Enormous amounts of information about stuttering are now available to anyone who can access the Internet. Compared to 20 years ago, people who stutter and their families can now make more informed choices about speech-language interventions, from a distance. Blogs and chat rooms provide opportunities for people who stutter to share their experiences from a distance and to support one another. New technologies are also being adopted into speech-language pathology practice and service delivery. Telehealth is an exciting development as it means that treatment can now be made available to many rural and remotely located people who previously did not have access to it. Possible future technological developments for speech-language pathology practice include Internet based treatments and the use of Virtual Reality. Having speech and CBT treatments for stuttering available on the Internet would greatly increase their accessibility. Second Life also has exciting possibilities for people who stutter. The reader will (1) explain how people who stutter and their families can get information about stuttering from the World Wide Web, (2) discuss how new technologies have been applied in speech-language pathology practice, and (3) summarize the principles and practice of telehealth delivery of services for people who stutter and their families. Copyright © 2011. Published by Elsevier Inc.

  13. Reprint of: technology and the evolution of clinical methods for stuttering.

    PubMed

    Packman, Ann; Meredith, Grant

    2011-09-01

    The World Wide Web (WWW) was 20 years old last year. Enormous amounts of information about stuttering are now available to anyone who can access the Internet. Compared to 20 years ago, people who stutter and their families can now make more informed choices about speech-language interventions, from a distance. Blogs and chat rooms provide opportunities for people who stutter to share their experiences from a distance and to support one another. New technologies are also being adopted into speech-language pathology practice and service delivery. Telehealth is an exciting development as it means that treatment can now be made available to many rural and remotely located people who previously did not have access to it. Possible future technological developments for speech-language pathology practice include Internet based treatments and the use of Virtual Reality. Having speech and CBT treatments for stuttering available on the Internet would greatly increase their accessibility. Second Life also has exciting possibilities for people who stutter. The reader will (1) explain how people who stutter and their families can get information about stuttering from the World Wide Web, (2) discuss how new technologies have been applied in speech-language pathology practice, and (3) summarize the principles and practice of telehealth delivery of services for people who stutter and their families. Copyright © 2011. Published by Elsevier Inc.

  14. Culturally diverse attitudes and beliefs of students majoring in speech-language pathology.

    PubMed

    Franca, Maria Claudia; Smith, Linda McCabe; Nichols, Jane Luanne; Balan, Dianna Santos

    Academic education in speech-language pathology should prepare students to provide professional services that mirror current knowledge, skills, and scope of practice in a pluralistic society. This study seeks to examine the impact of speech-language pathology (SLP) students prior multicultural experiences and previous formal education on attitudes and beliefs toward language diversity. A survey to investigate SLP students attitudes toward language diversity was applied. After the research study and instructions to complete the consent form questionnaire was presented by a research assistant, an announcement was given by a graduate student who speaks English as a second language with an accent. The participants then completed a questionnaire containing questions related to attitudes about the presentation of the announcement in particular and toward language diversity in general. Responses suggested a relationship between self-reported cultural bias and ability to concentrate on speech with an accent, and the extent of interaction with individuals from a cultural and linguistic diverse (CLD) background. Additional outcomes revealed that cultural bias may be predicted by factors related to amount of CLD exposure. Results of this study indicated critical areas that need to be considered when developing curricula in speech-language pathology programs. The results will be useful in determining procedures applicable in larger investigations, and encourage future research on attitudes and beliefs toward aspects of cultural diversity.

  15. Parent and child ratings of satisfaction with speech and facial appearance in Flemish pre-pubescent boys and girls with unilateral cleft lip and palate.

    PubMed

    Van Lierde, K M; Dhaeseleer, E; Luyten, A; Van De Woestijne, K; Vermeersch, H; Roche, N

    2012-02-01

    The purpose of this controlled study is to determine satisfaction with speech and facial appearance in Flemish pre-pubescent children with unilateral cleft lip and palate. Forty-three subjects with unilateral cleft lip and palate and 43 age and gender matched controls participated in this study. The Cleft Evaluation Profile was used to assess the perceived satisfaction for individual features related to cleft care. Both the cleft palate subjects and their parents were satisfied with the speech and facial appearance. The Pearson χ(2) test revealed significant difference between the cleft palate and the control group regarding hearing, nasal aesthetics and function, and the appearance of the lip. An in depth analysis of well specified speech characteristics revealed that children with clefts and their parents significantly more often reported the presence of an articulation, voice and resonance disorder and experienced /s/ /r/ /t/ and /d/ as the most difficult consonants. To what extent the incorporation of specific motor oriented oral speech techniques regarding the realisation of specific consonants, attention to vocal and ear care, and the recommendation of secondary velopharyngeal surgery, with the incorporation of primary correction of the cleft nose deformity simultaneously with primary lip closure, will aid these patients are future research subjects. Copyright © 2011 International Association of Oral and Maxillofacial Surgeons. Published by Elsevier Ltd. All rights reserved.

  16. Foucault's "fearless speech" and the transformation and mentoring of medical students

    PubMed Central

    Papadimos, Thomas J; Murray, Stuart J

    2008-01-01

    In his six 1983 lectures published under the title, Fearless Speech (2001), Michel Foucault developed the theme of free speech and its relation to frankness, truth-telling, criticism, and duty. Derived from the ancient Greek word parrhesia, Foucault's analysis of free speech is relevant to the mentoring of medical students. This is especially true given the educational and social need to transform future physicians into able citizens who practice a fearless freedom of expression on behalf of their patients, the public, the medical profession, and themselves in the public and political arena. In this paper, we argue that Foucault's understanding of free speech, or parrhesia, should be read as an ethical response to the American Medical Association's recent educational effort, Initiative to Transform Medical Education (ITME): Recommendations for change in the system of medical education (2007). In this document, the American Medical Association identifies gaps in medical education, emphasizing the need to enhance health system safety and quality, to improve education in training institutions, and to address the inadequacy of physician preparedness in new content areas. These gaps, and their relationship to the ITME goal of promoting excellence in patient care by implementing reform in the US system of medical education, call for a serious consideration and use of Foucault's parrhesia in the way that medical students are trained and mentored. PMID:18419809

  17. A comparison of sensory-motor activity during speech in first and second languages.

    PubMed

    Simmonds, Anna J; Wise, Richard J S; Dhanjal, Novraj S; Leech, Robert

    2011-07-01

    A foreign language (L2) learned after childhood results in an accent. This functional neuroimaging study investigated speech in L2 as a sensory-motor skill. The hypothesis was that there would be an altered response in auditory and somatosensory association cortex, specifically the planum temporale and parietal operculum, respectively, when speaking in L2 relative to L1, independent of rate of speaking. These regions were selected for three reasons. First, an influential computational model proposes that these cortices integrate predictive feedforward and postarticulatory sensory feedback signals during articulation. Second, these adjacent regions (known as Spt) have been identified as a "sensory-motor interface" for speech production. Third, probabilistic anatomical atlases exist for these regions, to ensure the analyses are confined to sensory-motor differences between L2 and L1. The study used functional magnetic resonance imaging (fMRI), and participants produced connected overt speech. The first hypothesis was that there would be greater activity in the planum temporale and the parietal operculum when subjects spoke in L2 compared with L1, one interpretation being that there is less efficient postarticulatory sensory monitoring when speaking in the less familiar L2. The second hypothesis was that this effect would be observed in both cerebral hemispheres. Although Spt is considered to be left-lateralized, this is based on studies of covert speech, whereas overt speech is accompanied by sensory feedback to bilateral auditory and somatosensory cortices. Both hypotheses were confirmed by the results. These findings provide the basis for future investigations of sensory-motor aspects of language learning using serial fMRI studies.

  18. Decoding Speech With Integrated Hybrid Signals Recorded From the Human Ventral Motor Cortex.

    PubMed

    Ibayashi, Kenji; Kunii, Naoto; Matsuo, Takeshi; Ishishita, Yohei; Shimada, Seijiro; Kawai, Kensuke; Saito, Nobuhito

    2018-01-01

    Restoration of speech communication for locked-in patients by means of brain computer interfaces (BCIs) is currently an important area of active research. Among the neural signals obtained from intracranial recordings, single/multi-unit activity (SUA/MUA), local field potential (LFP), and electrocorticography (ECoG) are good candidates for an input signal for BCIs. However, the question of which signal or which combination of the three signal modalities is best suited for decoding speech production remains unverified. In order to record SUA, LFP, and ECoG simultaneously from a highly localized area of human ventral sensorimotor cortex (vSMC), we fabricated an electrode the size of which was 7 by 13 mm containing sparsely arranged microneedle and conventional macro contacts. We determined which signal modality is the most capable of decoding speech production, and tested if the combination of these signals could improve the decoding accuracy of spoken phonemes. Feature vectors were constructed from spike frequency obtained from SUAs and event-related spectral perturbation derived from ECoG and LFP signals, then input to the decoder. The results showed that the decoding accuracy for five spoken vowels was highest when features from multiple signals were combined and optimized for each subject, and reached 59% when averaged across all six subjects. This result suggests that multi-scale signals convey complementary information for speech articulation. The current study demonstrated that simultaneous recording of multi-scale neuronal activities could raise decoding accuracy even though the recording area is limited to a small portion of cortex, which is advantageous for future implementation of speech-assisting BCIs.

  19. Decoding Speech With Integrated Hybrid Signals Recorded From the Human Ventral Motor Cortex

    PubMed Central

    Ibayashi, Kenji; Kunii, Naoto; Matsuo, Takeshi; Ishishita, Yohei; Shimada, Seijiro; Kawai, Kensuke; Saito, Nobuhito

    2018-01-01

    Restoration of speech communication for locked-in patients by means of brain computer interfaces (BCIs) is currently an important area of active research. Among the neural signals obtained from intracranial recordings, single/multi-unit activity (SUA/MUA), local field potential (LFP), and electrocorticography (ECoG) are good candidates for an input signal for BCIs. However, the question of which signal or which combination of the three signal modalities is best suited for decoding speech production remains unverified. In order to record SUA, LFP, and ECoG simultaneously from a highly localized area of human ventral sensorimotor cortex (vSMC), we fabricated an electrode the size of which was 7 by 13 mm containing sparsely arranged microneedle and conventional macro contacts. We determined which signal modality is the most capable of decoding speech production, and tested if the combination of these signals could improve the decoding accuracy of spoken phonemes. Feature vectors were constructed from spike frequency obtained from SUAs and event-related spectral perturbation derived from ECoG and LFP signals, then input to the decoder. The results showed that the decoding accuracy for five spoken vowels was highest when features from multiple signals were combined and optimized for each subject, and reached 59% when averaged across all six subjects. This result suggests that multi-scale signals convey complementary information for speech articulation. The current study demonstrated that simultaneous recording of multi-scale neuronal activities could raise decoding accuracy even though the recording area is limited to a small portion of cortex, which is advantageous for future implementation of speech-assisting BCIs. PMID:29674950

  20. Cognitive, Linguistic, and Motor Abilities in a Multigenerational Family with Childhood Apraxia of Speech.

    PubMed

    Carrigg, Bronwyn; Parry, Louise; Baker, Elise; Shriberg, Lawrence D; Ballard, Kirrie J

    2016-10-05

    This study describes the phenotype in a large family with a strong, multigenerational history of severe speech sound disorder (SSD) persisting into adolescence and adulthood in approximately half the cases. Aims were to determine whether a core phenotype, broader than speech, separated persistent from resolved SSD cases; and to ascertain the uniqueness of the phenotype relative to published cases. Eleven members of the PM family (9-55 years) were assessed across cognitive, language, literacy, speech, phonological processing, numeracy, and motor domains. Between group comparisons were made using the Mann-Whitney U-test (p < 0.01). Participant performances were compared to normative data using standardized tests and to the limited published data on persistent SSD phenotypes. Significant group differences were evident on multiple speech, language, literacy, phonological processing, and verbal intellect measures without any overlapping scores. Persistent cases performed within the impaired range on multiple measures. Phonological memory impairment and subtle literacy weakness were present in resolved SSD cases. A core phenotype distinguished persistent from resolved SSD cases that was characterized by a multiple verbal trait disorder, including Childhood Apraxia of Speech. Several phenotypic differences differentiated the persistent SSD phenotype in the PM family from the few previously reported studies of large families with SSD, including the absence of comorbid dysarthria and marked orofacial apraxia. This study highlights how comprehensive phenotyping can advance the behavioral study of disorders, in addition to forming a solid basis for future genetic and neural studies. © The Author 2016. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  1. Test–retest repeatability of human speech biomarkers from static and real-time dynamic magnetic resonance imaging

    PubMed Central

    Töger, Johannes; Sorensen, Tanner; Somandepalli, Krishna; Toutios, Asterios; Lingala, Sajan Goud; Narayanan, Shrikanth; Nayak, Krishna

    2017-01-01

    Static anatomical and real-time dynamic magnetic resonance imaging (RT-MRI) of the upper airway is a valuable method for studying speech production in research and clinical settings. The test–retest repeatability of quantitative imaging biomarkers is an important parameter, since it limits the effect sizes and intragroup differences that can be studied. Therefore, this study aims to present a framework for determining the test–retest repeatability of quantitative speech biomarkers from static MRI and RT-MRI, and apply the framework to healthy volunteers. Subjects (n = 8, 4 females, 4 males) are imaged in two scans on the same day, including static images and dynamic RT-MRI of speech tasks. The inter-study agreement is quantified using intraclass correlation coefficient (ICC) and mean within-subject standard deviation (σe). Inter-study agreement is strong to very strong for static measures (ICC: min/median/max 0.71/0.89/0.98, σe: 0.90/2.20/6.72 mm), poor to strong for dynamic RT-MRI measures of articulator motion range (ICC: 0.26/0.75/0.90, σe: 1.6/2.5/3.6 mm), and poor to very strong for velocities (ICC: 0.21/0.56/0.93, σe: 2.2/4.4/16.7 cm/s). In conclusion, this study characterizes repeatability of static and dynamic MRI-derived speech biomarkers using state-of-the-art imaging. The introduced framework can be used to guide future development of speech biomarkers. Test–retest MRI data are provided free for research use. PMID:28599561

  2. Behavior Assessment Battery: A Pilot Study of the Affective, Behavioral, and Cognitive Correlates Surrounding Spasmodic Dysphonia.

    PubMed

    Vanryckeghem, Martine; Hoffman Ruddy, Bari; Lehman, Jeffrey

    2016-01-01

    This study investigates if adults with adductor spasmodic dysphonia (ADSD) report to experience anxiety and voice problems in particular situations, indicate the presence of negative speech-associated attitude, and/or the use of coping behaviors, by means of the Behavior Assessment Battery (BAB) modified for voice. Thirty-two participants with ADSD and 32 adults without a voice disorder participated in this study. Each person completed four different BAB-Voice subtests. These standardized self-report tests are adaptations of the original BAB for people who stutter and explore an individual's speech-related belief, negative emotional reaction to and speech problems in particular speech situations, and the use of concomitant behaviors. Individuals with spasmodic dysphonia (SD) scored statistically significantly higher compared to typical speakers on all BAB subtests, indicating that individuals with SD report being significantly more anxious and experiencing significantly more voice problems in particular speech circumstances. They also reported a significant amount of negative speech-associated attitude and the use of a significant number of coping behaviors. Internal reliability was good for three of the four BAB subtests. The BAB is capable of reflecting the dimensions that surround the disorder of SD. The self-report measures have the potential to augment the observations made by the clinician and may lead to a more diverse and all-encompassing therapy for the person suffering from SD. Future research with a revised version of the BAB-Voice will continue to explore the validity, reliability, and replicability of the initial data. Published by Elsevier Inc.

  3. Masked speech perception across the adult lifespan: Impact of age and hearing impairment.

    PubMed

    Goossens, Tine; Vercammen, Charlotte; Wouters, Jan; van Wieringen, Astrid

    2017-02-01

    As people grow older, speech perception difficulties become highly prevalent, especially in noisy listening situations. Moreover, it is assumed that speech intelligibility is more affected in the event of background noises that induce a higher cognitive load, i.e., noises that result in informational versus energetic masking. There is ample evidence showing that speech perception problems in aging persons are partly due to hearing impairment and partly due to age-related declines in cognition and suprathreshold auditory processing. In order to develop effective rehabilitation strategies, it is indispensable to know how these different degrading factors act upon speech perception. This implies disentangling effects of hearing impairment versus age and examining the interplay between both factors in different background noises of everyday settings. To that end, we investigated open-set sentence identification in six participant groups: a young (20-30 years), middle-aged (50-60 years), and older cohort (70-80 years), each including persons who had normal audiometric thresholds up to at least 4 kHz, on the one hand, and persons who were diagnosed with elevated audiometric thresholds, on the other hand. All participants were screened for (mild) cognitive impairment. We applied stationary and amplitude modulated speech-weighted noise, which are two types of energetic maskers, and unintelligible speech, which causes informational masking in addition to energetic masking. By means of these different background noises, we could look into speech perception performance in listening situations with a low and high cognitive load, respectively. Our results indicate that, even when audiometric thresholds are within normal limits up to 4 kHz, irrespective of threshold elevations at higher frequencies, and there is no indication of even mild cognitive impairment, masked speech perception declines by middle age and decreases further on to older age. The impact of hearing impairment is as detrimental for young and middle-aged as it is for older adults. When the background noise becomes cognitively more demanding, there is a larger decline in speech perception, due to age or hearing impairment. Hearing impairment seems to be the main factor underlying speech perception problems in background noises that cause energetic masking. However, in the event of informational masking, which induces a higher cognitive load, age appears to explain a significant part of the communicative impairment as well. We suggest that the degrading effect of age is mediated by deficiencies in temporal processing and central executive functions. This study may contribute to the improvement of auditory rehabilitation programs aiming to prevent aging persons from missing out on conversations, which, in turn, will improve their quality of life. Copyright © 2016 Elsevier B.V. All rights reserved.

  4. Acoustically-Evoked Auditory Change Complex in Children with Auditory Neuropathy Spectrum Disorder: A Potential Objective Tool for Identifying Cochlear Implant Candidates

    PubMed Central

    He, Shuman; Grose, John H.; Teagle, Holly F.B.; Woodard, Jennifer; Park, Lisa R.; Hatch, Debora R.; Roush, Patricia; Buchman, Craig A.

    2014-01-01

    Objective: The overall aim of the study was to evaluate the feasibility of using electrophysiological measures of the auditory change complex (ACC) to identify candidates for cochlear implantation in children with auditory neuropathy spectrum disorder (ANSD). In order to achieve this overall aim, this study 1) assessed the feasibility of measuring the ACC evoked by temporal gaps in a group of children with ANSD across a wide age range; and 2) investigated the association between gap detection thresholds (GDTs) measured by the ACC recordings and open-set speech-perception performance in these subjects. Design: Nineteen children with bilateral ANSD ranging in age between 1.9 to 14.9 yrs (mean: 7.8 yrs) participated in this study. Electrophysiological recordings of the auditory event-related potential (ERP), including the onset ERP response and the ACC, were completed in all subjects and open-set speech perception was evaluated for a subgroup of sixteen subjects. For the ERP recordings, the stimulus was a Gaussian noise presented through ER-3A insert earphones to the test ear. Two stimulation conditions were used. In the “control condition,” the stimulus was an 800-ms Gaussian noise. In the “gapped condition”, the stimuli were two noise segments, each being 400 ms in duration, separated by one of five gaps (i.e. 5, 10, 20, 50, or 100 ms). The inter-stimulation interval was 1200 ms. The aided open-set speech perception ability was assessed using the Phonetically Balanced Kindergarten (PBK) word lists presented at 60 dB SPL using recorded testing material in a sound booth. For speech perception tests, subjects wore their hearing aids at the settings recommended by their clinical audiologists. For a subgroup of five subjects, psychophysical gap detection thresholds for the Gaussian noise were also assessed using a three-interval, three-alternative forced-choice procedure. Results: Responses evoked by the onset of the Gaussian noise (i.e. onset responses) were recorded in all stimulation conditions from all subjects tested in this study. The presence/absence, peak latency and amplitude, and response width of the onset response did not correlate with aided PBK word scores. The objective GDTs measured with the ACC recordings from seventeen subjects ranged from 10 to 100 ms. The ACC was not recorded from two subjects for any gap durations tested in this study. There was a robust negative correlation between objective GDTs and aided PBK word scores. In general, subjects with prolonged objective GDTs showed low aided PBK word scores. GDTs measured using electrophysiological recordings of the ACC correlated well with those measured using psychophysical procedures in four of five subjects who were evaluated using both procedures. Conclusions: The clinical application of the onset response in predicting open-set speech-perception ability is relatively limited in children with ANSD. The ACC recordings can be used to objectively evaluate temporal resolution abilities in children with ANSD having no severe comorbidities, and who are older than 1.9 years. The ACC can potentially be used as an objective tool to identify poor performers among children with ANSD using properly fit amplification, and who are thus, cochlear implant candidates. PMID:25422994

  5. A study of voice production characteristics of astronuat speech during Apollo 11 for speaker modeling in space.

    PubMed

    Yu, Chengzhu; Hansen, John H L

    2017-03-01

    Human physiology has evolved to accommodate environmental conditions, including temperature, pressure, and air chemistry unique to Earth. However, the environment in space varies significantly compared to that on Earth and, therefore, variability is expected in astronauts' speech production mechanism. In this study, the variations of astronaut voice characteristics during the NASA Apollo 11 mission are analyzed. Specifically, acoustical features such as fundamental frequency and phoneme formant structure that are closely related to the speech production system are studied. For a further understanding of astronauts' vocal tract spectrum variation in space, a maximum likelihood frequency warping based analysis is proposed to detect the vocal tract spectrum displacement during space conditions. The results from fundamental frequency, formant structure, as well as vocal spectrum displacement indicate that astronauts change their speech production mechanism when in space. Moreover, the experimental results for astronaut voice identification tasks indicate that current speaker recognition solutions are highly vulnerable to astronaut voice production variations in space conditions. Future recommendations from this study suggest that successful applications of speaker recognition during extended space missions require robust speaker modeling techniques that could effectively adapt to voice production variation caused by diverse space conditions.

  6. Decoding the attended speech stream with multi-channel EEG: implications for online, daily-life applications

    NASA Astrophysics Data System (ADS)

    Mirkovic, Bojana; Debener, Stefan; Jaeger, Manuela; De Vos, Maarten

    2015-08-01

    Objective. Recent studies have provided evidence that temporal envelope driven speech decoding from high-density electroencephalography (EEG) and magnetoencephalography recordings can identify the attended speech stream in a multi-speaker scenario. The present work replicated the previous high density EEG study and investigated the necessary technical requirements for practical attended speech decoding with EEG. Approach. Twelve normal hearing participants attended to one out of two simultaneously presented audiobook stories, while high density EEG was recorded. An offline iterative procedure eliminating those channels contributing the least to decoding provided insight into the necessary channel number and optimal cross-subject channel configuration. Aiming towards the future goal of near real-time classification with an individually trained decoder, the minimum duration of training data necessary for successful classification was determined by using a chronological cross-validation approach. Main results. Close replication of the previously reported results confirmed the method robustness. Decoder performance remained stable from 96 channels down to 25. Furthermore, for less than 15 min of training data, the subject-independent (pre-trained) decoder performed better than an individually trained decoder did. Significance. Our study complements previous research and provides information suggesting that efficient low-density EEG online decoding is within reach.

  7. Educating speech-language pathologists for the 21st century: course design considerations for a distance education Master of Speech Pathology program.

    PubMed

    McCormack, Jane; Easton, Catherine; Morkel-Kingsbury, Lenni

    2014-01-01

    The landscape of tertiary education is changing. Developments in information and communications technology have created new ways of engaging with subject material and supporting students on their learning journeys. Therefore, it is timely to reconsider and re-imagine the education of speech-language pathology (SLP) students within this new learning space. In this paper, we outline the design of a new Master of Speech Pathology course being offered by distance education at Charles Sturt University (CSU) in Australia. We discuss the catalyst for the course and the commitments of the SLP team at CSU, then describe the curriculum design process, focusing on the pedagogical approach and the learning and teaching strategies utilised in the course delivery. We explain how the learning and teaching strategies have been selected to support students' online learning experience and enable greater interaction between students and the subject material, with students and subject experts, and among student groups. Finally, we highlight some of the challenges in designing and delivering a distance education SLP program and identify future directions for educating students in an online world. © 2015 S. Karger AG, Basel.

  8. Analysis of 3-D Tongue Motion From Tagged and Cine Magnetic Resonance Images

    PubMed Central

    Woo, Jonghye; Lee, Junghoon; Murano, Emi Z.; Stone, Maureen; Prince, Jerry L.

    2016-01-01

    Purpose Measuring tongue deformation and internal muscle motion during speech has been a challenging task because the tongue deforms in 3 dimensions, contains interdigitated muscles, and is largely hidden within the vocal tract. In this article, a new method is proposed to analyze tagged and cine magnetic resonance images of the tongue during speech in order to estimate 3-dimensional tissue displacement and deformation over time. Method The method involves computing 2-dimensional motion components using a standard tag-processing method called harmonic phase, constructing superresolution tongue volumes using cine magnetic resonance images, segmenting the tongue region using a random-walker algorithm, and estimating 3-dimensional tongue motion using an incompressible deformation estimation algorithm. Results Evaluation of the method is presented with a control group and a group of people who had received a glossectomy carrying out a speech task. A 2-step principal-components analysis is then used to reveal the unique motion patterns of the subjects. Azimuth motion angles and motion on the mirrored hemi-tongues are analyzed. Conclusion Tests of the method with a various collection of subjects show its capability of capturing patient motion patterns and indicate its potential value in future speech studies. PMID:27295428

  9. A survey of music therapists' work with speech-language pathologists and experiences with augmentative and alternative communication.

    PubMed

    McCarthy, John; Geist, Kamile; Zojwala, Rashida; Schock, Molly Z

    2008-01-01

    Although music therapists may work with a variety of professionals in interdisciplinary teams, there is a lack of information about the specific nature of their work with speech-language pathologists (SLPs). Using an Internet-based tool, Board Certified Music Therapists (n = 1834, 1675 deliverable) were surveyed regarding their work with speech-language pathologists and experiences with augmentative and alternative communication (AAC). Specifically, participants were asked about: (a) demographics; (b) populations worked with professionally; (c) past and present work with speech-language pathologists; (d) goals addressed; (e) benefits and challenges encountered; and (f) work with AAC. Responses (N = 847) indicated the majority of participants (73.6%) had worked with SLPs at some point in various roles and in various settings. Fewer participants reported currently working with SLPs (42.8%), although 50.1% reported currently working with someone requiring some form of AAC. Participants reported a mean level of expertise with AAC of 3.9 on a scale of 1-7. Sharing knowledge was noted as a top benefit of working with SLPs, while scheduling was reported as the most frequent challenge. Other benefits and challenges as well as future research directions are discussed.

  10. Re-Establishing Broca's Initial Findings

    ERIC Educational Resources Information Center

    Richardson, Jessica D.; Fillmore, Paul; Rorden, Chris; LaPointe, Leonard L.; Fridriksson, Julius

    2012-01-01

    The importance of the left inferior pre-frontal cortex (LIPC) for speech production was first popularized by Paul Broca, providing a cornerstone of behavioral neurology and laying the foundation for future research examining brain-behavior relationships. Although Broca's findings were rigorously challenged, comprehensive contradictory evidence was…

  11. Computer content analysis of schizophrenic speech: a preliminary report.

    PubMed

    Tucker, G J; Rosenberg, S D

    1975-06-01

    Computer analysis significantly differtiated the thermatic content of the free speech of 10 schizophrenic patients from that of 10 nonschizophrenic patients and from the content of transcripts of dream material from 10 normal subjects. Schizophrenic patients used the thematic categories in factor 1 (the "schizophrenic factor") 3 times more frequently than the nonschizophrenics and 10 times more frequently than the normal subjects (p smaller than 01). In general, the language content of the schizophrenic patient mirrored an almost agitated attempt to locate oneself in time and space and to defend against internal discomfort and confusion. The authors discuss the implications of this study for future research.

  12. Scientific bases of human-machine communication by voice.

    PubMed Central

    Schafer, R W

    1995-01-01

    The scientific bases for human-machine communication by voice are in the fields of psychology, linguistics, acoustics, signal processing, computer science, and integrated circuit technology. The purpose of this paper is to highlight the basic scientific and technological issues in human-machine communication by voice and to point out areas of future research opportunity. The discussion is organized around the following major issues in implementing human-machine voice communication systems: (i) hardware/software implementation of the system, (ii) speech synthesis for voice output, (iii) speech recognition and understanding for voice input, and (iv) usability factors related to how humans interact with machines. PMID:7479802

  13. An assessment of the information-seeking abilities and needs of practicing speech-language pathologists

    PubMed Central

    Nail-Chiwetalu, Barbara; Bernstein Ratner, Nan

    2007-01-01

    Objectives: This study assessed the information-seeking practices and needs of speech-language pathologists (SLPs). Improved understanding of these needs can inform librarians and educators to better prepare students in principles and methods of evidence-based practice (EBP) and, through continuing education (CE), promote the integration of EBP into clinical practice of SLPs. Methods: A 16-question survey was mailed to 1,000 certified speech-language pathologists in the United States. Results: Two hundred and eight usable surveys were returned for a response rate of 21%. For clinical questions, SLPs most often consulted with a colleague, participated in CE activities, and searched the open Internet. Few respondents relied on scholarly journal articles for assistance with clinical cases. The most prominent barriers to finding appropriate information were time and knowledge of where and how to find relevant information. Few reported having information literacy instruction by a librarian. Discussion: If EBP is to become a viable practice in clinical decision making, there appears to be a tremendous need for information literacy instruction in the university curriculum, as well as through CE activities for currently practicing SLPs. Given respondents' reported lack of time and limited access to full-text journals containing evidence relevant to clinical practice, the field of speech-language pathology will need to generate readily accessible clinical summaries of research evidence through meta-analyses, systematic reviews, and clinical practice guidelines. PMID:17443251

  14. The Mental Capacity Bill 2004: its impact on district nurse practice.

    PubMed

    Griffith, Richard; Tengnah, Cassam

    2005-01-01

    The government made clear its intention to reform the law relating to mentally incapable adults in the Queen's speech at the state opening of Parliament in November 2004. In this article Richard Griffith and Cassam Tengnah outline the main provisions of the Mental Capacity Bill and assess its likely impact on district nurses who care for adults with incapacity.

  15. Acoustic Correlates of Inflectional Morphology in the Speech of Children with Specific Language Impairment and Their Typically Developing Peers

    ERIC Educational Resources Information Center

    Owen, Amanda J.; Goffman, Lisa

    2007-01-01

    The development of the use of the third-person singular -s in open syllable verbs in children with specific language impairment (SLI) and their typically developing peers was examined. Verbs that included overt productions of the third-person singular -s morpheme (e.g. "Bobby plays ball everyday;" "Bear laughs when mommy buys…

  16. When Poetry became Ethnography and Other Flying Pig Tales in Honor of Dell Hymes

    ERIC Educational Resources Information Center

    Cahnmann-Taylor, Melisa

    2011-01-01

    Cahnmann-Taylor remembers her first encounter with Dell Hymes at an open mic event at the annual meeting of the American Anthropological Association. She puzzles his complex stance on the role ethnographic poems might play in one's ethnographic project. In Dell Hymes's honor, she shares a poetic rendering of a speech event from her bilingual…

  17. National Workshop on the Preparation of Literacy Follow-Up Materials in Malaysia (Penang, Malaysia, December 9-19, 1990). Report.

    ERIC Educational Resources Information Center

    Asian Cultural Centre for UNESCO, Tokyo (Japan).

    This report contains materials from a workshop aimed at the production and usage of video programs in the field of functional literacy, especially as literacy follow-up materials. Chapter I contains the proceedings and consists of an introduction, objective, and summaries of opening speeches, presentations, and group work. Chapter II provides…

  18. The Prosodic Licensing of Coda Consonants in Early Speech: Interactions with Vowel Length

    ERIC Educational Resources Information Center

    Miles, Kelly; Yuen, Ivan; Cox, Felicity; Demuth, Katherine

    2016-01-01

    English has a word-minimality requirement that all open-class lexical items must contain at least two moras of structure, forming a bimoraic foot (Hayes, 1995).Thus, a word with either a long vowel, or a short vowel and a coda consonant, satisfies this requirement. This raises the question of when and how young children might learn this…

  19. Using Vision and Speech Features for Automated Prediction of Performance Metrics in Multimodal Dialogs. Research Report. ETS RR-17-20

    ERIC Educational Resources Information Center

    Ramanarayanan, Vikram; Lange, Patrick; Evanini, Keelan; Molloy, Hillary; Tsuprun, Eugene; Qian, Yao; Suendermann-Oeft, David

    2017-01-01

    Predicting and analyzing multimodal dialog user experience (UX) metrics, such as overall call experience, caller engagement, and latency, among other metrics, in an ongoing manner is important for evaluating such systems. We investigate automated prediction of multiple such metrics collected from crowdsourced interactions with an open-source,…

  20. The Speech Focus Position Effect on Jaw-Finger Coordination in a Pointing Task

    ERIC Educational Resources Information Center

    Rochet-Capellan, Amelie; Laboissiere, Rafael; Galvan, Arturo; Schwartz, Jean-Luc

    2008-01-01

    Purpose: This article investigates jaw-finger coordination in a task involving pointing to a target while naming it with a 'CVCV (e.g., /'papa/) versus CV'CV (e.g., /pa'pa/) word. According to the authors' working hypothesis, the pointing apex (gesture extremum) would be synchronized with the apex of the jaw-opening gesture corresponding to the…

Top