Sample records for direct voice input

  1. Voice and gesture-based 3D multimedia presentation tool

    NASA Astrophysics Data System (ADS)

    Fukutake, Hiromichi; Akazawa, Yoshiaki; Okada, Yoshihiro

    2007-09-01

    This paper proposes a 3D multimedia presentation tool that allows the user to manipulate intuitively only through the voice input and the gesture input without using a standard keyboard or a mouse device. The authors developed this system as a presentation tool to be used in a presentation room equipped a large screen like an exhibition room in a museum because, in such a presentation environment, it is better to use voice commands and the gesture pointing input rather than using a keyboard or a mouse device. This system was developed using IntelligentBox, which is a component-based 3D graphics software development system. IntelligentBox has already provided various types of 3D visible, reactive functional components called boxes, e.g., a voice input component and various multimedia handling components. IntelligentBox also provides a dynamic data linkage mechanism called slot-connection that allows the user to develop 3D graphics applications by combining already existing boxes through direct manipulations on a computer screen. Using IntelligentBox, the 3D multimedia presentation tool proposed in this paper was also developed as combined components only through direct manipulations on a computer screen. The authors have already proposed a 3D multimedia presentation tool using a stage metaphor and its voice input interface. This time, we extended the system to make it accept the user gesture input besides voice commands. This paper explains details of the proposed 3D multimedia presentation tool and especially describes its component-based voice and gesture input interfaces.

  2. Voice recognition products-an occupational risk for users with ULDs?

    PubMed

    Williams, N R

    2003-10-01

    Voice recognition systems (VRS) allow speech to be converted both directly into text-which appears on the screen of a computer-and to direct equipment to perform specific functions. Suggested applications are many and varied, including increasing efficiency in the reporting of radiographs, allowing directed surgery and enabling individuals with upper limb disorders (ULDs) who cannot use other input devices, such as keyboards and mice, to carry out word processing and other activities. Aim This paper describes four cases of vocal dysfunction related to the use of such software, which have been identified from the database of the Voice and Speech Laboratory of the Massachusetts Eye and Ear infirmary (MEEI). The database was searched using key words 'voice recognition' and four cases were identified from a total of 4800. In all cases, the VRS was supplied to assist individuals with ULDs who could not use conventional input devices. Case reports illustrate time of onset and symptoms experienced. The cases illustrate the need for risk assessment and consideration of the ergonomic aspects of voice use prior to such adaptations being used, particularly in those who already experience work-related ULDs.

  3. Smartphone Text Input Method Performance, Usability, and Preference With Younger and Older Adults.

    PubMed

    Smith, Amanda L; Chaparro, Barbara S

    2015-09-01

    User performance, perceived usability, and preference for five smartphone text input methods were compared with younger and older novice adults. Smartphones are used for a variety of functions other than phone calls, including text messaging, e-mail, and web browsing. Research comparing performance with methods of text input on smartphones reveals a high degree of variability in reported measures, procedures, and results. This study reports on a direct comparison of five of the most common input methods among a population of younger and older adults, who had no experience with any of the methods. Fifty adults (25 younger, 18-35 years; 25 older, 60-84 years) completed a text entry task using five text input methods (physical Qwerty, onscreen Qwerty, tracing, handwriting, and voice). Entry and error rates, perceived usability, and preference were recorded. Both age groups input text equally fast using voice input, but older adults were slower than younger adults using all other methods. Both age groups had low error rates when using physical Qwerty and voice, but older adults committed more errors with the other three methods. Both younger and older adults preferred voice and physical Qwerty input to the remaining methods. Handwriting consistently performed the worst and was rated lowest by both groups. Voice and physical Qwerty input methods proved to be the most effective for both younger and older adults, and handwriting input was the least effective overall. These findings have implications to the design of future smartphone text input methods and devices, particularly for older adults. © 2015, Human Factors and Ergonomics Society.

  4. Speech versus manual control of camera functions during a telerobotic task

    NASA Technical Reports Server (NTRS)

    Bierschwale, John M.; Sampaio, Carlos E.; Stuart, Mark A.; Smith, Randy L.

    1989-01-01

    Voice input for control of camera functions was investigated in this study. Objective were to (1) assess the feasibility of a voice-commanded camera control system, and (2) identify factors that differ between voice and manual control of camera functions. Subjects participated in a remote manipulation task that required extensive camera-aided viewing. Each subject was exposed to two conditions, voice and manual input, with a counterbalanced administration order. Voice input was found to be significantly slower than manual input for this task. However, in terms of remote manipulator performance errors and subject preference, there was no difference between modalities. Voice control of continuous camera functions is not recommended. It is believed that the use of voice input for discrete functions, such as multiplexing or camera switching, could aid performance. Hybrid mixes of voice and manual input may provide the best use of both modalities. This report contributes to a better understanding of the issues that affect the design of an efficient human/telerobot interface.

  5. Using Natural Language to Enhance Mission Effectiveness

    NASA Technical Reports Server (NTRS)

    Trujillo, Anna C.; Meszaros, Erica

    2016-01-01

    The availability of highly capable, yet relatively cheap, unmanned aerial vehicles (UAVs) is opening up new areas of use for hobbyists and for professional-related activities. The driving function of this research is allowing a non-UAV pilot, an operator, to define and manage a mission. This paper describes the preliminary usability measures of an interface that allows an operator to define the mission using speech to make inputs. An experiment was conducted to begin to enumerate the efficacy and user acceptance of using voice commands to define a multi-UAV mission and to provide high-level vehicle control commands such as "takeoff." The primary independent variable was input type - voice or mouse. The primary dependent variables consisted of the correctness of the mission parameter inputs and the time needed to make all inputs. Other dependent variables included NASA-TLX workload ratings and subjective ratings on a final questionnaire. The experiment required each subject to fill in an online form that contained comparable required information that would be needed for a package dispatcher to deliver packages. For each run, subjects typed in a simple numeric code for the package code. They then defined the initial starting position, the delivery location, and the return location using either pull-down menus or voice input. Voice input was accomplished using CMU Sphinx4-5prealpha for speech recognition. They then inputted the length of the package. These were the option fields. The subject had the system "Calculate Trajectory" and then "Takeoff" once the trajectory was calculated. Later, the subject used "Land" to finish the run. After the voice and mouse input blocked runs, subjects completed a NASA-TLX. At the conclusion of all runs, subjects completed a questionnaire asking them about their experience in inputting the mission parameters, and starting and stopping the mission using mouse and voice input. In general, the usability of voice commands is acceptable. With a relatively well-defined and simple vocabulary, the operator can input the vast majority of the mission parameters using simple, intuitive voice commands. However, voice input may be more applicable to initial mission specification rather than for critical commands such as the need to land immediately due to time and feedback constraints. It would also be convenient to retrieve relevant mission information using voice input. Therefore, further on-going research is looking at using intent from operator utterances to provide the relevant mission information to the operator. The information displayed will be inferred from the operator's utterances just before key phrases are spoken. Linguistic analysis of the context of verbal communication provides insight into the intended meaning of commonly heard phrases such as "What's it doing now?" Analyzing the semantic sphere surrounding these common phrases enables us to predict the operator's intent and supply the operator's desired information to the interface. This paper also describes preliminary investigations into the generation of the semantic space of UAV operation and the success at providing information to the interface based on the operator's utterances.

  6. Human voice quality measurement in noisy environments.

    PubMed

    Ueng, Shyh-Kuang; Luo, Cheng-Ming; Tsai, Tsung-Yu; Yeh, Hsuan-Chen

    2015-01-01

    Computerized acoustic voice measurement is essential for the diagnosis of vocal pathologies. Previous studies showed that ambient noises have significant influences on the accuracy of voice quality assessment. This paper presents a voice quality assessment system that can accurately measure qualities of voice signals, even though the input voice data are contaminated by low-frequency noises. The ambient noises in our living rooms and laboratories are collected and the frequencies of these noises are analyzed. Based on the analysis, a filter is designed to reduce noise level of the input voice signal. Then, improved numerical algorithms are employed to extract voice parameters from the voice signal to reveal the health of the voice signal. Compared with MDVP and Praat, the proposed method outperforms these two widely used programs in measuring fundamental frequency and harmonic-to-noise ratio, and its performance is comparable to these two famous programs in computing jitter and shimmer. The proposed voice quality assessment method is resistant to low-frequency noises and it can measure human voice quality in environments filled with noises from air-conditioners, ceiling fans and cooling fans of computers.

  7. The recognition of female voice based on voice registers in singing techniques in real-time using hankel transform method and macdonald function

    NASA Astrophysics Data System (ADS)

    Meiyanti, R.; Subandi, A.; Fuqara, N.; Budiman, M. A.; Siahaan, A. P. U.

    2018-03-01

    A singer doesn’t just recite the lyrics of a song, but also with the use of particular sound techniques to make it more beautiful. In the singing technique, more female have a diverse sound registers than male. There are so many registers of the human voice, but the voice registers used while singing, among others, Chest Voice, Head Voice, Falsetto, and Vocal fry. Research of speech recognition based on the female’s voice registers in singing technique is built using Borland Delphi 7.0. Speech recognition process performed by the input recorded voice samples and also in real time. Voice input will result in weight energy values based on calculations using Hankel Transformation method and Macdonald Functions. The results showed that the accuracy of the system depends on the accuracy of sound engineering that trained and tested, and obtained an average percentage of the successful introduction of the voice registers record reached 48.75 percent, while the average percentage of the successful introduction of the voice registers in real time to reach 57 percent.

  8. Speech versus manual control of camera functions during a telerobotic task

    NASA Technical Reports Server (NTRS)

    Bierschwale, John M.; Sampaio, Carlos E.; Stuart, Mark A.; Smith, Randy L.

    1993-01-01

    This investigation has evaluated the voice-commanded camera control concept. For this particular task, total voice control of continuous and discrete camera functions was significantly slower than manual control. There was no significant difference between voice and manual input for several types of errors. There was not a clear trend in subjective preference of camera command input modality. Task performance, in terms of both accuracy and speed, was very similar across both levels of experience.

  9. Enhanced Living by Assessing Voice Pathology Using a Co-Occurrence Matrix

    PubMed Central

    Muhammad, Ghulam; Alhamid, Mohammed F.; Hossain, M. Shamim; Almogren, Ahmad S.; Vasilakos, Athanasios V.

    2017-01-01

    A large number of the population around the world suffers from various disabilities. Disabilities affect not only children but also adults of different professions. Smart technology can assist the disabled population and lead to a comfortable life in an enhanced living environment (ELE). In this paper, we propose an effective voice pathology assessment system that works in a smart home framework. The proposed system takes input from various sensors, and processes the acquired voice signals and electroglottography (EGG) signals. Co-occurrence matrices in different directions and neighborhoods from the spectrograms of these signals were obtained. Several features such as energy, entropy, contrast, and homogeneity from these matrices were calculated and fed into a Gaussian mixture model-based classifier. Experiments were performed with a publicly available database, namely, the Saarbrucken voice database. The results demonstrate the feasibility of the proposed system in light of its high accuracy and speed. The proposed system can be extended to assess other disabilities in an ELE. PMID:28146069

  10. Enhanced Living by Assessing Voice Pathology Using a Co-Occurrence Matrix.

    PubMed

    Muhammad, Ghulam; Alhamid, Mohammed F; Hossain, M Shamim; Almogren, Ahmad S; Vasilakos, Athanasios V

    2017-01-29

    A large number of the population around the world suffers from various disabilities. Disabilities affect not only children but also adults of different professions. Smart technology can assist the disabled population and lead to a comfortable life in an enhanced living environment (ELE). In this paper, we propose an effective voice pathology assessment system that works in a smart home framework. The proposed system takes input from various sensors, and processes the acquired voice signals and electroglottography (EGG) signals. Co-occurrence matrices in different directions and neighborhoods from the spectrograms of these signals were obtained. Several features such as energy, entropy, contrast, and homogeneity from these matrices were calculated and fed into a Gaussian mixture model-based classifier. Experiments were performed with a publicly available database, namely, the Saarbrucken voice database. The results demonstrate the feasibility of the proposed system in light of its high accuracy and speed. The proposed system can be extended to assess other disabilities in an ELE.

  11. Voice Response Systems Technology.

    ERIC Educational Resources Information Center

    Gerald, Jeanette

    1984-01-01

    Examines two methods of generating synthetic speech in voice response systems, which allow computers to communicate in human terms (speech), using human interface devices (ears): phoneme and reconstructed voice systems. Considerations prior to implementation, current and potential applications, glossary, directory, and introduction to Input Output…

  12. Speaking Math--A Voice Input, Speech Output Calculator for Students with Visual Impairments

    ERIC Educational Resources Information Center

    Bouck, Emily C.; Flanagan, Sara; Joshi, Gauri S.; Sheikh, Waseem; Schleppenbach, Dave

    2011-01-01

    This project explored a newly developed computer-based voice input, speech output (VISO) calculator. Three high school students with visual impairments educated at a state school for the blind and visually impaired participated in the study. The time they took to complete assessments and the average number of attempts per problem were recorded…

  13. The Influence of Visual Feedback and Register Changes on Sign Language Production: A Kinematic Study with Deaf Signers

    ERIC Educational Resources Information Center

    Emmorey, Karen; Gertsberg, Nelly; Korpics, Franco; Wright, Charles E.

    2009-01-01

    Speakers monitor their speech output by listening to their own voice. However, signers do not look directly at their hands and cannot see their own face. We investigated the importance of a visual perceptual loop for sign language monitoring by examining whether changes in visual input alter sign production. Deaf signers produced American Sign…

  14. Robotics control using isolated word recognition of voice input

    NASA Technical Reports Server (NTRS)

    Weiner, J. M.

    1977-01-01

    A speech input/output system is presented that can be used to communicate with a task oriented system. Human speech commands and synthesized voice output extend conventional information exchange capabilities between man and machine by utilizing audio input and output channels. The speech input facility is comprised of a hardware feature extractor and a microprocessor implemented isolated word or phrase recognition system. The recognizer offers a medium sized (100 commands), syntactically constrained vocabulary, and exhibits close to real time performance. The major portion of the recognition processing required is accomplished through software, minimizing the complexity of the hardware feature extractor.

  15. A Phenomenological Study: Perceptions of Student Voice on Academic Success

    ERIC Educational Resources Information Center

    Marberry, Tammie

    2013-01-01

    The purpose of this qualitative, phenomenological study was to explore rural high school graduates', teachers', and administrators' perceptions of student voice on academic success. This study was designed to examine the following three questions: What were the common beliefs regarding opportunities for input, or student voice, on the educational…

  16. Driving While Interacting With Google Glass: Investigating the Combined Effect of Head-Up Display and Hands-Free Input on Driving Safety and Multitask Performance.

    PubMed

    Tippey, Kathryn G; Sivaraj, Elayaraj; Ferris, Thomas K

    2017-06-01

    This study evaluated the individual and combined effects of voice (vs. manual) input and head-up (vs. head-down) display in a driving and device interaction task. Advances in wearable technology offer new possibilities for in-vehicle interaction but also present new challenges for managing driver attention and regulating device usage in vehicles. This research investigated how driving performance is affected by interface characteristics of devices used for concurrent secondary tasks. A positive impact on driving performance was expected when devices included voice-to-text functionality (reducing demand for visual and manual resources) and a head-up display (HUD) (supporting greater visibility of the driving environment). Driver behavior and performance was compared in a texting-while-driving task set during a driving simulation. The texting task was completed with and without voice-to-text using a smartphone and with voice-to-text using Google Glass's HUD. Driving task performance degraded with the addition of the secondary texting task. However, voice-to-text input supported relatively better performance in both driving and texting tasks compared to using manual entry. HUD functionality further improved driving performance compared to conditions using a smartphone and often was not significantly worse than performance without the texting task. This study suggests that despite the performance costs of texting-while-driving, voice input methods improve performance over manual entry, and head-up displays may further extend those performance benefits. This study can inform designers and potential users of wearable technologies as well as policymakers tasked with regulating the use of these technologies while driving.

  17. Evolving Spiking Neural Networks for Recognition of Aged Voices.

    PubMed

    Silva, Marco; Vellasco, Marley M B R; Cataldo, Edson

    2017-01-01

    The aging of the voice, known as presbyphonia, is a natural process that can cause great change in vocal quality of the individual. This is a relevant problem to those people who use their voices professionally, and its early identification can help determine a suitable treatment to avoid its progress or even to eliminate the problem. This work focuses on the development of a new model for the identification of aging voices (independently of their chronological age), using as input attributes parameters extracted from the voice and glottal signals. The proposed model, named Quantum binary-real evolving Spiking Neural Network (QbrSNN), is based on spiking neural networks (SNNs), with an unsupervised training algorithm, and a Quantum-Inspired Evolutionary Algorithm that automatically determines the most relevant attributes and the optimal parameters that configure the SNN. The QbrSNN model was evaluated in a database composed of 120 records, containing samples from three groups of speakers. The results obtained indicate that the proposed model provides better accuracy than other approaches, with fewer input attributes. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  18. Noise Source Visualization Using a Digital Voice Recorder and Low-Cost Sensors

    PubMed Central

    Cho, Yong Thung

    2018-01-01

    Accurate sound visualization of noise sources is required for optimal noise control. Typically, noise measurement systems require microphones, an analog-digital converter, cables, a data acquisition system, etc., which may not be affordable for potential users. Also, many such systems are not highly portable and may not be convenient for travel. Handheld personal electronic devices such as smartphones and digital voice recorders with relatively lower costs and higher performance have become widely available recently. Even though such devices are highly portable, directly implementing them for noise measurement may lead to erroneous results since such equipment was originally designed for voice recording. In this study, external microphones were connected to a digital voice recorder to conduct measurements and the input received was processed for noise visualization. In this way, a low cost, compact sound visualization system was designed and introduced to visualize two actual noise sources for verification with different characteristics: an enclosed loud speaker and a small air compressor. Reasonable accuracy of noise visualization for these two sources was shown over a relatively wide frequency range. This very affordable and compact sound visualization system can be used for many actual noise visualization applications in addition to educational purposes. PMID:29614038

  19. Interventions for preventing voice disorders in adults.

    PubMed

    Ruotsalainen, J H; Sellman, J; Lehto, L; Jauhiainen, M; Verbeek, J H

    2007-10-17

    Poor voice quality due to a voice disorder can lead to a reduced quality of life. In occupations where voice use is substantial it can lead to periods of absence from work. To evaluate the effectiveness of interventions to prevent voice disorders in adults. We searched MEDLINE (PubMed, 1950 to 2006), EMBASE (1974 to 2006), CENTRAL (The Cochrane Library, Issue 2 2006), CINAHL (1983 to 2006), PsychINFO (1967 to 2006), Science Citation Index (1986 to 2006) and the Occupational Health databases OSH-ROM (to 2006). The date of the last search was 05/04/06. Randomised controlled clinical trials (RCTs) of interventions evaluating the effectiveness of treatments to prevent voice disorders in adults. For work-directed interventions interrupted time series and prospective cohort studies were also eligible. Two authors independently extracted data and assessed trial quality. Meta-analysis was performed where appropriate. We identified two randomised controlled trials including a total of 53 participants in intervention groups and 43 controls. One study was conducted with teachers and the other with student teachers. Both trials were poor quality. Interventions were grouped into 1) direct voice training, 2) indirect voice training and 3) direct and indirect voice training combined.1) Direct voice training: One study did not find a significant decrease of the Voice Handicap Index for direct voice training compared to no intervention.2) Indirect voice training: One study did not find a significant decrease of the Voice Handicap Index for indirect voice training when compared to no intervention.3) Direct and indirect voice training combined: One study did not find a decrease of the Voice Handicap Index for direct and indirect voice training combined when compared to no intervention. The same study did however find an improvement in maximum phonation time (Mean Difference -3.18 sec; 95 % CI -4.43 to -1.93) for direct and indirect voice training combined when compared to no intervention. No work-directed studies were found. None of the studies found evaluated the effectiveness of prevention in terms of sick leave or number of diagnosed voice disorders. We found no evidence that either direct or indirect voice training or the two combined are effective in improving self-reported vocal functioning when compared to no intervention. The current practice of giving training to at-risk populations for preventing the development of voice disorders is therefore not supported by definitive evidence of effectiveness. Larger and methodologically better trials are needed with outcome measures that better reflect the aims of interventions.

  20. Scientific bases of human-machine communication by voice.

    PubMed Central

    Schafer, R W

    1995-01-01

    The scientific bases for human-machine communication by voice are in the fields of psychology, linguistics, acoustics, signal processing, computer science, and integrated circuit technology. The purpose of this paper is to highlight the basic scientific and technological issues in human-machine communication by voice and to point out areas of future research opportunity. The discussion is organized around the following major issues in implementing human-machine voice communication systems: (i) hardware/software implementation of the system, (ii) speech synthesis for voice output, (iii) speech recognition and understanding for voice input, and (iv) usability factors related to how humans interact with machines. PMID:7479802

  1. HELP: Handheld Emergency Logistics Program for Generating Structured Requests for Resources in Stressful Conditions

    DTIC Science & Technology

    2014-09-01

    Redesign .................................122 d. Screen 10/Final Review Redesign ........................................123 F. TEST SET- UP INITIAL TEST...user with a chance to review his or her inputs and send the request by his or her preferred method (digital or voice). The screen breaks down the line...user with a chance to review his or her inputs and send the request by his or her preferred method (digital or voice). The screen breaks down the

  2. Motorcycle Start-stop System based on Intelligent Biometric Voice Recognition

    NASA Astrophysics Data System (ADS)

    Winda, A.; E Byan, W. R.; Sofyan; Armansyah; Zariantin, D. L.; Josep, B. G.

    2017-03-01

    Current mechanical key in the motorcycle is prone to bulgary, being stolen or misplaced. Intelligent biometric voice recognition as means to replace this mechanism is proposed as an alternative. The proposed system will decide whether the voice is belong to the user or not and the word utter by the user is ‘On’ or ‘Off’. The decision voice will be sent to Arduino in order to start or stop the engine. The recorded voice is processed in order to get some features which later be used as input to the proposed system. The Mel-Frequency Ceptral Coefficient (MFCC) is adopted as a feature extraction technique. The extracted feature is the used as input to the SVM-based identifier. Experimental results confirm the effectiveness of the proposed intelligent voice recognition and word recognition system. It show that the proposed method produces a good training and testing accuracy, 99.31% and 99.43%, respectively. Moreover, the proposed system shows the performance of false rejection rate (FRR) and false acceptance rate (FAR) accuracy of 0.18% and 17.58%, respectively. In the intelligent word recognition shows that the training and testing accuracy are 100% and 96.3%, respectively.

  3. A voice-input voice-output communication aid for people with severe speech impairment.

    PubMed

    Hawley, Mark S; Cunningham, Stuart P; Green, Phil D; Enderby, Pam; Palmer, Rebecca; Sehgal, Siddharth; O'Neill, Peter

    2013-01-01

    A new form of augmentative and alternative communication (AAC) device for people with severe speech impairment-the voice-input voice-output communication aid (VIVOCA)-is described. The VIVOCA recognizes the disordered speech of the user and builds messages, which are converted into synthetic speech. System development was carried out employing user-centered design and development methods, which identified and refined key requirements for the device. A novel methodology for building small vocabulary, speaker-dependent automatic speech recognizers with reduced amounts of training data, was applied. Experiments showed that this method is successful in generating good recognition performance (mean accuracy 96%) on highly disordered speech, even when recognition perplexity is increased. The selected message-building technique traded off various factors including speed of message construction and range of available message outputs. The VIVOCA was evaluated in a field trial by individuals with moderate to severe dysarthria and confirmed that they can make use of the device to produce intelligible speech output from disordered speech input. The trial highlighted some issues which limit the performance and usability of the device when applied in real usage situations, with mean recognition accuracy of 67% in these circumstances. These limitations will be addressed in future work.

  4. Audio signal processor

    NASA Technical Reports Server (NTRS)

    Hymer, R. L.

    1970-01-01

    System provides automatic volume control for an audio amplifier or a voice communication system without introducing noise surges during pauses in the input, and without losing the initial signal when the input resumes.

  5. Community Agency Voice and Benefit in Service-Learning

    ERIC Educational Resources Information Center

    Miron, Devi; Moely, Barbara E.

    2006-01-01

    Supervisors from 40 community agencies working with a university-based service-learning program were interviewed regarding the extent of their input in service-learning program planning and implementation "(Agency Voice), Interpersonal Relations" with service-learning students, "Perceived Benefit" of the service-learning…

  6. Three input concepts for flight crew interaction with information presented on a large-screen electronic cockpit display

    NASA Technical Reports Server (NTRS)

    Jones, Denise R.

    1990-01-01

    A piloted simulation study was conducted comparing three different input methods for interfacing to a large-screen, multiwindow, whole-flight-deck display for management of transport aircraft systems. The thumball concept utilized a miniature trackball embedded in a conventional side-arm controller. The touch screen concept provided data entry through a capacitive touch screen. The voice concept utilized a speech recognition system with input through a head-worn microphone. No single input concept emerged as the most desirable method of interacting with the display. Subjective results, however, indicate that the voice concept was the most preferred method of data entry and had the most potential for future applications. The objective results indicate that, overall, the touch screen concept was the most effective input method. There was also significant differences between the time required to perform specific tasks and the input concept employed, with each concept providing better performance relative to a specific task. These results suggest that a system combining all three input concepts might provide the most effective method of interaction.

  7. Natural asynchronies in audiovisual communication signals regulate neuronal multisensory interactions in voice-sensitive cortex.

    PubMed

    Perrodin, Catherine; Kayser, Christoph; Logothetis, Nikos K; Petkov, Christopher I

    2015-01-06

    When social animals communicate, the onset of informative content in one modality varies considerably relative to the other, such as when visual orofacial movements precede a vocalization. These naturally occurring asynchronies do not disrupt intelligibility or perceptual coherence. However, they occur on time scales where they likely affect integrative neuronal activity in ways that have remained unclear, especially for hierarchically downstream regions in which neurons exhibit temporally imprecise but highly selective responses to communication signals. To address this, we exploited naturally occurring face- and voice-onset asynchronies in primate vocalizations. Using these as stimuli we recorded cortical oscillations and neuronal spiking responses from functional MRI (fMRI)-localized voice-sensitive cortex in the anterior temporal lobe of macaques. We show that the onset of the visual face stimulus resets the phase of low-frequency oscillations, and that the face-voice asynchrony affects the prominence of two key types of neuronal multisensory responses: enhancement or suppression. Our findings show a three-way association between temporal delays in audiovisual communication signals, phase-resetting of ongoing oscillations, and the sign of multisensory responses. The results reveal how natural onset asynchronies in cross-sensory inputs regulate network oscillations and neuronal excitability in the voice-sensitive cortex of macaques, a suggested animal model for human voice areas. These findings also advance predictions on the impact of multisensory input on neuronal processes in face areas and other brain regions.

  8. The eye-voice span during reading aloud

    PubMed Central

    Laubrock, Jochen; Kliegl, Reinhold

    2015-01-01

    Although eye movements during reading are modulated by cognitive processing demands, they also reflect visual sampling of the input, and possibly preparation of output for speech or the inner voice. By simultaneously recording eye movements and the voice during reading aloud, we obtained an output measure that constrains the length of time spent on cognitive processing. Here we investigate the dynamics of the eye-voice span (EVS), the distance between eye and voice. We show that the EVS is regulated immediately during fixation of a word by either increasing fixation duration or programming a regressive eye movement against the reading direction. EVS size at the beginning of a fixation was positively correlated with the likelihood of regressions and refixations. Regression probability was further increased if the EVS was still large at the end of a fixation: if adjustment of fixation duration did not sufficiently reduce the EVS during a fixation, then a regression rather than a refixation followed with high probability. We further show that the EVS can help understand cognitive influences on fixation duration during reading: in mixed model analyses, the EVS was a stronger predictor of fixation durations than either word frequency or word length. The EVS modulated the influence of several other predictors on single fixation durations (SFDs). For example, word-N frequency effects were larger with a large EVS, especially when word N-1 frequency was low. Finally, a comparison of SFDs during oral and silent reading showed that reading is governed by similar principles in both reading modes, although EVS maintenance and articulatory processing also cause some differences. In summary, the EVS is regulated by adjusting fixation duration and/or by programming a regressive eye movement when the EVS gets too large. Overall, the EVS appears to be directly related to updating of the working memory buffer during reading. PMID:26441800

  9. Effects of emotional and perceptual-motor stress on a voice recognition system's accuracy: An applied investigation

    NASA Astrophysics Data System (ADS)

    Poock, G. K.; Martin, B. J.

    1984-02-01

    This was an applied investigation examining the ability of a speech recognition system to recognize speakers' inputs when the speakers were under different stress levels. Subjects were asked to speak to a voice recognition system under three conditions: (1) normal office environment, (2) emotional stress, and (3) perceptual-motor stress. Results indicate a definite relationship between voice recognition system performance and the type of low stress reference patterns used to achieve recognition.

  10. Literature review of voice recognition and generation technology for Army helicopter applications

    NASA Astrophysics Data System (ADS)

    Christ, K. A.

    1984-08-01

    This report is a literature review on the topics of voice recognition and generation. Areas covered are: manual versus vocal data input, vocabulary, stress and workload, noise, protective masks, feedback, and voice warning systems. Results of the studies presented in this report indicate that voice data entry has less of an impact on a pilot's flight performance, during low-level flying and other difficult missions, than manual data entry. However, the stress resulting from such missions may cause the pilot's voice to change, reducing the recognition accuracy of the system. The noise present in helicopter cockpits also causes the recognition accuracy to decrease. Noise-cancelling devices are being developed and improved upon to increase the recognition performance in noisy environments. Future research in the fields of voice recognition and generation should be conducted in the areas of stress and workload, vocabulary, and the types of voice generation best suited for the helicopter cockpit. Also, specific tasks should be studied to determine whether voice recognition and generation can be effectively applied.

  11. Natural asynchronies in audiovisual communication signals regulate neuronal multisensory interactions in voice-sensitive cortex

    PubMed Central

    Perrodin, Catherine; Kayser, Christoph; Logothetis, Nikos K.; Petkov, Christopher I.

    2015-01-01

    When social animals communicate, the onset of informative content in one modality varies considerably relative to the other, such as when visual orofacial movements precede a vocalization. These naturally occurring asynchronies do not disrupt intelligibility or perceptual coherence. However, they occur on time scales where they likely affect integrative neuronal activity in ways that have remained unclear, especially for hierarchically downstream regions in which neurons exhibit temporally imprecise but highly selective responses to communication signals. To address this, we exploited naturally occurring face- and voice-onset asynchronies in primate vocalizations. Using these as stimuli we recorded cortical oscillations and neuronal spiking responses from functional MRI (fMRI)-localized voice-sensitive cortex in the anterior temporal lobe of macaques. We show that the onset of the visual face stimulus resets the phase of low-frequency oscillations, and that the face–voice asynchrony affects the prominence of two key types of neuronal multisensory responses: enhancement or suppression. Our findings show a three-way association between temporal delays in audiovisual communication signals, phase-resetting of ongoing oscillations, and the sign of multisensory responses. The results reveal how natural onset asynchronies in cross-sensory inputs regulate network oscillations and neuronal excitability in the voice-sensitive cortex of macaques, a suggested animal model for human voice areas. These findings also advance predictions on the impact of multisensory input on neuronal processes in face areas and other brain regions. PMID:25535356

  12. Interface Anywhere: Development of a Voice and Gesture System for Spaceflight Operations

    NASA Technical Reports Server (NTRS)

    Thompson, Shelby; Haddock, Maxwell; Overland, David

    2013-01-01

    The Interface Anywhere Project was funded through Innovation Charge Account (ICA) at NASA JSC in the Fall of 2012. The project was collaboration between human factors and engineering to explore the possibility of designing an interface to control basic habitat operations through gesture and voice control; (a) Current interfaces require the users to be physically near an input device in order to interact with the system; and (b) By using voice and gesture commands, the user is able to interact with the system anywhere they want within the work environment.

  13. The Effect of Processing Instruction and Dictogloss Tasks on Acquisition of the English Passive Voice

    ERIC Educational Resources Information Center

    Qin, Jingjing

    2008-01-01

    This study was intended to compare processing instruction (VanPatten, 1993, 1996, 2000), an input-based focus on form technique, to dictogloss tasks, an output-oriented focus-on-form type of instruction to assess their effects in helping beginning-EFL (English as a Foreign Language) learners acquire the simple English passive voice. Two intact…

  14. The instrumental phase of the voice program at the Utrecht school of acting.

    PubMed

    Schrama, Els

    2008-01-01

    What skills does a performer need in order to be able to say their lines on stage? What is the input of an actor to be audible and have a lively voice filled with imagination? To train the professional performer, we need to know the purpose and the way to arrive there. Copyright 2008 S. Karger AG, Basel.

  15. Wavelet-based associative memory

    NASA Astrophysics Data System (ADS)

    Jones, Katharine J.

    2004-04-01

    Faces provide important characteristics of a person"s identification. In security checks, face recognition still remains the method in continuous use despite other approaches (i.e. fingerprints, voice recognition, pupil contraction, DNA scanners). With an associative memory, the output data is recalled directly using the input data. This can be achieved with a Nonlinear Holographic Associative Memory (NHAM). This approach can also distinguish between strongly correlated images and images that are partially or totally enclosed by others. Adaptive wavelet lifting has been used for Content-Based Image Retrieval. In this paper, adaptive wavelet lifting will be applied to face recognition to achieve an associative memory.

  16. A Voice and Mouse Input Interface for 3D Virtual Environments

    NASA Technical Reports Server (NTRS)

    Kao, David L.; Bryson, Steve T.

    2003-01-01

    There have been many successful stories on how 3D input devices can be fully integrated into an immersive virtual environment. Electromagnetic trackers, optical trackers, gloves, and flying mice are just some of these input devices. Though we can use existing 3D input devices that are commonly used for VR applications, there are several factors that prevent us from choosing these input devices for our applications. One main factor is that most of these tracking devices are not suitable for prolonged use due to human fatigue associated with using them. A second factor is that many of them would occupy additional office space. Another factor is that many of the 3D input devices are expensive due to the unusual hardware that are required. For our VR applications, we want a user interface that would work naturally with standard equipment. In this paper, we demonstrate applications or our proposed muitimodal interface using a 3D dome display. We also show that effective data analysis can be achieved while the scientists view their data rendered inside the dome display and perform user interactions simply using the mouse and voice input. Though the sphere coordinate grid seems to be ideal for interaction using a 3D dome display, we can also use other non-spherical grids as well.

  17. Crossmodal adaptation in right posterior superior temporal sulcus during face-voice emotional integration.

    PubMed

    Watson, Rebecca; Latinus, Marianne; Noguchi, Takao; Garrod, Oliver; Crabbe, Frances; Belin, Pascal

    2014-05-14

    The integration of emotional information from the face and voice of other persons is known to be mediated by a number of "multisensory" cerebral regions, such as the right posterior superior temporal sulcus (pSTS). However, whether multimodal integration in these regions is attributable to interleaved populations of unisensory neurons responding to face or voice or rather by multimodal neurons receiving input from the two modalities is not fully clear. Here, we examine this question using functional magnetic resonance adaptation and dynamic audiovisual stimuli in which emotional information was manipulated parametrically and independently in the face and voice via morphing between angry and happy expressions. Healthy human adult subjects were scanned while performing a happy/angry emotion categorization task on a series of such stimuli included in a fast event-related, continuous carryover design. Subjects integrated both face and voice information when categorizing emotion-although there was a greater weighting of face information-and showed behavioral adaptation effects both within and across modality. Adaptation also occurred at the neural level: in addition to modality-specific adaptation in visual and auditory cortices, we observed for the first time a crossmodal adaptation effect. Specifically, fMRI signal in the right pSTS was reduced in response to a stimulus in which facial emotion was similar to the vocal emotion of the preceding stimulus. These results suggest that the integration of emotional information from face and voice in the pSTS involves a detectable proportion of bimodal neurons that combine inputs from visual and auditory cortices. Copyright © 2014 the authors 0270-6474/14/346813-09$15.00/0.

  18. Crossmodal Adaptation in Right Posterior Superior Temporal Sulcus during Face–Voice Emotional Integration

    PubMed Central

    Latinus, Marianne; Noguchi, Takao; Garrod, Oliver; Crabbe, Frances; Belin, Pascal

    2014-01-01

    The integration of emotional information from the face and voice of other persons is known to be mediated by a number of “multisensory” cerebral regions, such as the right posterior superior temporal sulcus (pSTS). However, whether multimodal integration in these regions is attributable to interleaved populations of unisensory neurons responding to face or voice or rather by multimodal neurons receiving input from the two modalities is not fully clear. Here, we examine this question using functional magnetic resonance adaptation and dynamic audiovisual stimuli in which emotional information was manipulated parametrically and independently in the face and voice via morphing between angry and happy expressions. Healthy human adult subjects were scanned while performing a happy/angry emotion categorization task on a series of such stimuli included in a fast event-related, continuous carryover design. Subjects integrated both face and voice information when categorizing emotion—although there was a greater weighting of face information—and showed behavioral adaptation effects both within and across modality. Adaptation also occurred at the neural level: in addition to modality-specific adaptation in visual and auditory cortices, we observed for the first time a crossmodal adaptation effect. Specifically, fMRI signal in the right pSTS was reduced in response to a stimulus in which facial emotion was similar to the vocal emotion of the preceding stimulus. These results suggest that the integration of emotional information from face and voice in the pSTS involves a detectable proportion of bimodal neurons that combine inputs from visual and auditory cortices. PMID:24828635

  19. A Development of a System Enables Character Input and PC Operation via Voice for a Physically Disabled Person with a Speech Impediment

    NASA Astrophysics Data System (ADS)

    Tanioka, Toshimasa; Egashira, Hiroyuki; Takata, Mayumi; Okazaki, Yasuhisa; Watanabe, Kenzi; Kondo, Hiroki

    We have designed and implemented a PC operation support system for a physically disabled person with a speech impediment via voice. Voice operation is an effective method for a physically disabled person with involuntary movement of the limbs and the head. We have applied a commercial speech recognition engine to develop our system for practical purposes. Adoption of a commercial engine reduces development cost and will contribute to make our system useful to another speech impediment people. We have customized commercial speech recognition engine so that it can recognize the utterance of a person with a speech impediment. We have restricted the words that the recognition engine recognizes and separated a target words from similar words in pronunciation to avoid misrecognition. Huge number of words registered in commercial speech recognition engines cause frequent misrecognition for speech impediments' utterance, because their utterance is not clear and unstable. We have solved this problem by narrowing the choice of input down in a small number and also by registering their ambiguous pronunciations in addition to the original ones. To realize all character inputs and all PC operation with a small number of words, we have designed multiple input modes with categorized dictionaries and have introduced two-step input in each mode except numeral input to enable correct operation with small number of words. The system we have developed is in practical level. The first author of this paper is physically disabled with a speech impediment. He has been able not only character input into PC but also to operate Windows system smoothly by using this system. He uses this system in his daily life. This paper is written by him with this system. At present, the speech recognition is customized to him. It is, however, possible to customize for other users by changing words and registering new pronunciation according to each user's utterance.

  20. The role of voice input for human-machine communication.

    PubMed Central

    Cohen, P R; Oviatt, S L

    1995-01-01

    Optimism is growing that the near future will witness rapid growth in human-computer interaction using voice. System prototypes have recently been built that demonstrate speaker-independent real-time speech recognition, and understanding of naturally spoken utterances with vocabularies of 1000 to 2000 words, and larger. Already, computer manufacturers are building speech recognition subsystems into their new product lines. However, before this technology can be broadly useful, a substantial knowledge base is needed about human spoken language and performance during computer-based spoken interaction. This paper reviews application areas in which spoken interaction can play a significant role, assesses potential benefits of spoken interaction with machines, and compares voice with other modalities of human-computer interaction. It also discusses information that will be needed to build a firm empirical foundation for the design of future spoken and multimodal interfaces. Finally, it argues for a more systematic and scientific approach to investigating spoken input and performance with future language technology. PMID:7479803

  1. Real-Time Reconfigurable Adaptive Speech Recognition Command and Control Apparatus and Method

    NASA Technical Reports Server (NTRS)

    Salazar, George A. (Inventor); Haynes, Dena S. (Inventor); Sommers, Marc J. (Inventor)

    1998-01-01

    An adaptive speech recognition and control system and method for controlling various mechanisms and systems in response to spoken instructions and in which spoken commands are effective to direct the system into appropriate memory nodes, and to respective appropriate memory templates corresponding to the voiced command is discussed. Spoken commands from any of a group of operators for which the system is trained may be identified, and voice templates are updated as required in response to changes in pronunciation and voice characteristics over time of any of the operators for which the system is trained. Provisions are made for both near-real-time retraining of the system with respect to individual terms which are determined not be positively identified, and for an overall system training and updating process in which recognition of each command and vocabulary term is checked, and in which the memory templates are retrained if necessary for respective commands or vocabulary terms with respect to an operator currently using the system. In one embodiment, the system includes input circuitry connected to a microphone and including signal processing and control sections for sensing the level of vocabulary recognition over a given period and, if recognition performance falls below a given level, processing audio-derived signals for enhancing recognition performance of the system.

  2. Intentional Voice Command Detection for Trigger-Free Speech Interface

    NASA Astrophysics Data System (ADS)

    Obuchi, Yasunari; Sumiyoshi, Takashi

    In this paper we introduce a new framework of audio processing, which is essential to achieve a trigger-free speech interface for home appliances. If the speech interface works continually in real environments, it must extract occasional voice commands and reject everything else. It is extremely important to reduce the number of false alarms because the number of irrelevant inputs is much larger than the number of voice commands even for heavy users of appliances. The framework, called Intentional Voice Command Detection, is based on voice activity detection, but enhanced by various speech/audio processing techniques such as emotion recognition. The effectiveness of the proposed framework is evaluated using a newly-collected large-scale corpus. The advantages of combining various features were tested and confirmed, and the simple LDA-based classifier demonstrated acceptable performance. The effectiveness of various methods of user adaptation is also discussed.

  3. Brain 'talks over' boring quotes: top-down activation of voice-selective areas while listening to monotonous direct speech quotations.

    PubMed

    Yao, Bo; Belin, Pascal; Scheepers, Christoph

    2012-04-15

    In human communication, direct speech (e.g., Mary said, "I'm hungry") is perceived as more vivid than indirect speech (e.g., Mary said that she was hungry). This vividness distinction has previously been found to underlie silent reading of quotations: Using functional magnetic resonance imaging (fMRI), we found that direct speech elicited higher brain activity in the temporal voice areas (TVA) of the auditory cortex than indirect speech, consistent with an "inner voice" experience in reading direct speech. Here we show that listening to monotonously spoken direct versus indirect speech quotations also engenders differential TVA activity. This suggests that individuals engage in top-down simulations or imagery of enriched supra-segmental acoustic representations while listening to monotonous direct speech. The findings shed new light on the acoustic nature of the "inner voice" in understanding direct speech. Copyright © 2012 Elsevier Inc. All rights reserved.

  4. Using Voice Coils to Actuate Modular Soft Robots: Wormbot, an Example.

    PubMed

    Nemitz, Markus P; Mihaylov, Pavel; Barraclough, Thomas W; Ross, Dylan; Stokes, Adam A

    2016-12-01

    In this study, we present a modular worm-like robot, which utilizes voice coils as a new paradigm in soft robot actuation. Drive electronics are incorporated into the actuators, providing a significant improvement in self-sufficiency when compared with existing soft robot actuation modes such as pneumatics or hydraulics. The body plan of this robot is inspired by the phylum Annelida and consists of three-dimensional printed voice coil actuators, which are connected by flexible silicone membranes. Each electromagnetic actuator engages with its neighbor to compress or extend the membrane of each segment, and the sequence in which they are actuated results in an earthworm-inspired peristaltic motion. We find that a minimum of three segments is required for locomotion, but due to our modular design, robots of any length can be quickly and easily assembled. In addition to actuation, voice coils provide audio input and output capabilities. We demonstrate transmission of data between segments by high-frequency carrier waves and, using a similar mechanism, we note that the passing of power between coupled coils in neighboring modules-or from an external power source-is also possible. Voice coils are a convenient multifunctional alternative to existing soft robot actuators. Their self-contained nature and ability to communicate with each other are ideal for modular robotics, and the additional functionality of sound input/output and power transfer will become increasingly useful as soft robots begin the transition from early proof-of-concept systems toward fully functional and highly integrated robotic systems.

  5. Psychological Therapies for Auditory Hallucinations (Voices): Current Status and Key Directions for Future Research

    PubMed Central

    Thomas, Neil; Hayward, Mark; Peters, Emmanuelle; van der Gaag, Mark; Bentall, Richard P.; Jenner, Jack; Strauss, Clara; Sommer, Iris E.; Johns, Louise C.; Varese, Filippo; García-Montes, José Manuel; Waters, Flavie; Dodgson, Guy; McCarthy-Jones, Simon

    2014-01-01

    This report from the International Consortium on Hallucinations Research considers the current status and future directions in research on psychological therapies targeting auditory hallucinations (hearing voices). Therapy approaches have evolved from behavioral and coping-focused interventions, through formulation-driven interventions using methods from cognitive therapy, to a number of contemporary developments. Recent developments include the application of acceptance- and mindfulness-based approaches, and consolidation of methods for working with connections between voices and views of self, others, relationships and personal history. In this article, we discuss the development of therapies for voices and review the empirical findings. This review shows that psychological therapies are broadly effective for people with positive symptoms, but that more research is required to understand the specific application of therapies to voices. Six key research directions are identified: (1) moving beyond the focus on overall efficacy to understand specific therapeutic processes targeting voices, (2) better targeting psychological processes associated with voices such as trauma, cognitive mechanisms, and personal recovery, (3) more focused measurement of the intended outcomes of therapy, (4) understanding individual differences among voice hearers, (5) extending beyond a focus on voices and schizophrenia into other populations and sensory modalities, and (6) shaping interventions for service implementation. PMID:24936081

  6. Crossmodal plasticity in the fusiform gyrus of late blind individuals during voice recognition.

    PubMed

    Hölig, Cordula; Föcker, Julia; Best, Anna; Röder, Brigitte; Büchel, Christian

    2014-12-01

    Blind individuals are trained in identifying other people through voices. In congenitally blind adults the anterior fusiform gyrus has been shown to be active during voice recognition. Such crossmodal changes have been associated with a superiority of blind adults in voice perception. The key question of the present functional magnetic resonance imaging (fMRI) study was whether visual deprivation that occurs in adulthood is followed by similar adaptive changes of the voice identification system. Late blind individuals and matched sighted participants were tested in a priming paradigm, in which two voice stimuli were subsequently presented. The prime (S1) and the target (S2) were either from the same speaker (person-congruent voices) or from two different speakers (person-incongruent voices). Participants had to classify the S2 as either coming from an old or a young person. Only in late blind but not in matched sighted controls, the activation in the anterior fusiform gyrus was modulated by voice identity: late blind volunteers showed an increase of the BOLD signal in response to person-incongruent compared with person-congruent trials. These results suggest that the fusiform gyrus adapts to input of a new modality even in the mature brain and thus demonstrate an adult type of crossmodal plasticity. Copyright © 2014 Elsevier Inc. All rights reserved.

  7. Application of AI techniques to a voice-actuated computer system for reconstructing and displaying magnetic resonance imaging data

    NASA Astrophysics Data System (ADS)

    Sherley, Patrick L.; Pujol, Alfonso, Jr.; Meadow, John S.

    1990-07-01

    To provide a means of rendering complex computer architectures languages and input/output modalities transparent to experienced and inexperienced users research is being conducted to develop a voice driven/voice response computer graphics imaging system. The system will be used for reconstructing and displaying computed tomography and magnetic resonance imaging scan data. In conjunction with this study an artificial intelligence (Al) control strategy was developed to interface the voice components and support software to the computer graphics functions implemented on the Sun Microsystems 4/280 color graphics workstation. Based on generated text and converted renditions of verbal utterances by the user the Al control strategy determines the user''s intent and develops and validates a plan. The program type and parameters within the plan are used as input to the graphics system for reconstructing and displaying medical image data corresponding to that perceived intent. If the plan is not valid the control strategy queries the user for additional information. The control strategy operates in a conversation mode and vocally provides system status reports. A detailed examination of the various AT techniques is presented with major emphasis being placed on their specific roles within the total control strategy structure. 1.

  8. Graphics with Special Interfaces for Disabled People.

    ERIC Educational Resources Information Center

    Tronconi, A.; And Others

    The paper describes new software and special input devices to allow physically impaired children to utilize the graphic capabilities of personal computers. Special input devices for computer graphics access--the voice recognition card, the single switch, or the mouse emulator--can be used either singly or in combination by the disabled to control…

  9. Smartphones Offer New Opportunities in Clinical Voice Research.

    PubMed

    Manfredi, C; Lebacq, J; Cantarella, G; Schoentgen, J; Orlandi, S; Bandini, A; DeJonckere, P H

    2017-01-01

    Smartphone technology provides new opportunities for recording standardized voice samples of patients and sending the files by e-mail to the voice laboratory. This drastically improves the collection of baseline data, as used in research on efficiency of voice treatments. However, the basic requirement is the suitability of smartphones for recording and digitizing pathologic voices (mainly characterized by period perturbations and noise) without significant distortion. In this experiment, two smartphones (a very inexpensive one and a high-level one) were tested and compared with direct microphone recordings in a soundproof room. The voice stimuli consisted in synthesized deviant voice samples (median of fundamental frequency: 120 and 200 Hz) with three levels of jitter and three levels of added noise. All voice samples were analyzed using PRAAT software. The results show high correlations between jitter, shimmer, and noise-to-harmonics ratio measured on the recordings via both smartphones, the microphone, and measured directly on the sound files from the synthesizer. Smartphones thus appear adequate for reliable recording and digitizing of pathologic voices. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  10. Indirect vs Direct Voice Therapy for Children With Vocal Nodules: A Randomized Clinical Trial.

    PubMed

    Hartnick, Christopher; Ballif, Catherine; De Guzman, Vanessa; Sataloff, Robert; Campisi, Paolo; Kerschner, Joseph; Shembel, Adrianna; Reda, Domenic; Shi, Helen; Sheryka Zacny, Elinore; Bunting, Glenn

    2018-02-01

    Benign vocal fold nodules affect 12% to 22% of the pediatric population, and 95% of otolaryngologists recommend voice therapy as treatment. However, no randomized clinical trials that we are aware of have shown its benefits. To determine the impact of voice therapy in children with vocal fold nodules according to pretherapy and posttherapy scores on the Pediatric Voice-Related Quality of Life (PVRQOL) survey; secondary objectives included changes in phonatory parameters. For this multicenter randomized clinical trial, 114 children ages 6 to 10 years with vocal fold nodules, PVRQOL scores less than 87.5, and dysphonia for longer than 12 weeks were recruited from outpatient voice and speech clinics. This age range was identified because these patients have not experienced pubertal changes of the larynx, tolerate stroboscopy, and cooperate with voice therapy. Participants were blinded to treatment arm. Participants received either indirect or direct therapy for 8 to 12 weeks. Indirect therapy focused on education and discussion of voice principles, while direct treatment used the stimulus, response, antecedent paradigm. The primary outcome measure was PVRQOL score change before and after treatment. Secondary phonatory measures were also compared. Overall, 114 children were recruited for study (mean [SD] age, 8 [1.4] years; 83 males [73%]); with 57 randomized to receive either indirect or direct therapy. Both direct and indirect therapy approaches showed significant differences in PVRQOL scores pretherapy to posttherapy. The mean increase in PVRQOL score for direct therapy was 19.2, and 14.7 for indirect therapy (difference, 4.5; 95.3% CI, -10.8 to 19.8). Of 44 participants in the direct therapy group, 27 (61%) achieved a clinically meaningful PVRQOL improvement, compared with 26 of 49 (53%) for indirect therapy (difference, 8%; 95% CI, -12 to 28). Post hoc stratification showed robust effects in the direct therapy group for older children (Cohen d = 0.50) and the latter two-thirds of participants (Cohen d = 0.46). Vocal fold nodules reduced in size in 31% (22 of 70) and completely resolved in 11% (8 of 70) of participants who consented to a second set of images after going through the recruitment process. Both direct and indirect voice therapy improved voice-related quality of life in children with vocal fold nodules, although there was no significant difference between approaches. Future studies may focus upon which voice therapy approaches are effective in treating age-defined populations. clinicaltrials.gov Identifier: NCT01255735.

  11. A unified coding strategy for processing faces and voices

    PubMed Central

    Yovel, Galit; Belin, Pascal

    2013-01-01

    Both faces and voices are rich in socially-relevant information, which humans are remarkably adept at extracting, including a person's identity, age, gender, affective state, personality, etc. Here, we review accumulating evidence from behavioral, neuropsychological, electrophysiological, and neuroimaging studies which suggest that the cognitive and neural processing mechanisms engaged by perceiving faces or voices are highly similar, despite the very different nature of their sensory input. The similarity between the two mechanisms likely facilitates the multi-modal integration of facial and vocal information during everyday social interactions. These findings emphasize a parsimonious principle of cerebral organization, where similar computational problems in different modalities are solved using similar solutions. PMID:23664703

  12. Second Report of the Multirate Processor (MRP) for Digital Voice Communications.

    DTIC Science & Technology

    1982-09-30

    machine are: * two arithmetic logic units (ALUs)-one for data processing, and the other for address generation, * two memorys -6144 words (70 bits per word...of program memory , and 6094 words (16 bits per word) of data memory , q * input/output through modem and teletype, -15 .9 S-;. KANG AND FRANSEN Table...provides a measure of intelligibility and allows one to evaluate the discriminability of six distinctive features: voicing, nasality, sustention

  13. Psychological therapies for auditory hallucinations (voices): current status and key directions for future research.

    PubMed

    Thomas, Neil; Hayward, Mark; Peters, Emmanuelle; van der Gaag, Mark; Bentall, Richard P; Jenner, Jack; Strauss, Clara; Sommer, Iris E; Johns, Louise C; Varese, Filippo; García-Montes, José Manuel; Waters, Flavie; Dodgson, Guy; McCarthy-Jones, Simon

    2014-07-01

    This report from the International Consortium on Hallucinations Research considers the current status and future directions in research on psychological therapies targeting auditory hallucinations (hearing voices). Therapy approaches have evolved from behavioral and coping-focused interventions, through formulation-driven interventions using methods from cognitive therapy, to a number of contemporary developments. Recent developments include the application of acceptance- and mindfulness-based approaches, and consolidation of methods for working with connections between voices and views of self, others, relationships and personal history. In this article, we discuss the development of therapies for voices and review the empirical findings. This review shows that psychological therapies are broadly effective for people with positive symptoms, but that more research is required to understand the specific application of therapies to voices. Six key research directions are identified: (1) moving beyond the focus on overall efficacy to understand specific therapeutic processes targeting voices, (2) better targeting psychological processes associated with voices such as trauma, cognitive mechanisms, and personal recovery, (3) more focused measurement of the intended outcomes of therapy, (4) understanding individual differences among voice hearers, (5) extending beyond a focus on voices and schizophrenia into other populations and sensory modalities, and (6) shaping interventions for service implementation. © The Author 2014. Published by Oxford University Press on behalf of the Maryland Psychiatric Research Center.

  14. Medications and Adverse Voice Effects.

    PubMed

    Nemr, Kátia; Di Carlos Silva, Ariana; Rodrigues, Danilo de Albuquerque; Zenari, Marcia Simões

    2017-08-16

    To identify the medications used by patients with dysphonia, describe the voice symptoms reported on initial speech-language pathology (SLP) examination, evaluate the possible direct and indirect effects of medications on voice production, and determine the association between direct and indirect adverse voice effects and self-reported voice symptoms, hydration and smoking habits, comorbidities, vocal assessment, and type and degree of dysphonia. This is a retrospective cross-sectional study. Fifty-five patients were evaluated and the vocal signs and symptoms indicated in the Dysphonia Risk Protocol were considered, as well as data on hydration, smoking and medication use. We analyzed the associations between type of side effect and self-reported vocal signs/symptoms, hydration, smoking, comorbidities, type of dysphonia, and auditory-perceptual and acoustic parameters. Sixty percent were women, the mean age was 51.8 years, 29 symptoms were reported on the screening, and 73 active ingredients were identified with 8.2% directly and 91.8% indirectly affecting vocal function. There were associations between the use of drugs with direct adverse voice effects, self-reported symptoms, general degree of vocal deviation, and pitch deviation. The symptoms of dry throat and shortness of breath were associated with the direct vocal side effect of the medicine, as well as the general degree of vocal deviation and the greater pitch deviation. Shortness of breath when speaking was also associated with the greatest degree of vocal deviation. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  15. Analysis and Classification of Voice Pathologies Using Glottal Signal Parameters.

    PubMed

    Forero M, Leonardo A; Kohler, Manoela; Vellasco, Marley M B R; Cataldo, Edson

    2016-09-01

    The classification of voice diseases has many applications in health, in diseases treatment, and in the design of new medical equipment for helping doctors in diagnosing pathologies related to the voice. This work uses the parameters of the glottal signal to help the identification of two types of voice disorders related to the pathologies of the vocal folds: nodule and unilateral paralysis. The parameters of the glottal signal are obtained through a known inverse filtering method, and they are used as inputs to an Artificial Neural Network, a Support Vector Machine, and also to a Hidden Markov Model, to obtain the classification, and to compare the results, of the voice signals into three different groups: speakers with nodule in the vocal folds; speakers with unilateral paralysis of the vocal folds; and speakers with normal voices, that is, without nodule or unilateral paralysis present in the vocal folds. The database is composed of 248 voice recordings (signals of vowels production) containing samples corresponding to the three groups mentioned. In this study, a larger database was used for the classification when compared with similar studies, and its classification rate is superior to other studies, reaching 97.2%. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  16. Using Voice Coils to Actuate Modular Soft Robots: Wormbot, an Example

    PubMed Central

    Nemitz, Markus P.; Mihaylov, Pavel; Barraclough, Thomas W.; Ross, Dylan

    2016-01-01

    Abstract In this study, we present a modular worm-like robot, which utilizes voice coils as a new paradigm in soft robot actuation. Drive electronics are incorporated into the actuators, providing a significant improvement in self-sufficiency when compared with existing soft robot actuation modes such as pneumatics or hydraulics. The body plan of this robot is inspired by the phylum Annelida and consists of three-dimensional printed voice coil actuators, which are connected by flexible silicone membranes. Each electromagnetic actuator engages with its neighbor to compress or extend the membrane of each segment, and the sequence in which they are actuated results in an earthworm-inspired peristaltic motion. We find that a minimum of three segments is required for locomotion, but due to our modular design, robots of any length can be quickly and easily assembled. In addition to actuation, voice coils provide audio input and output capabilities. We demonstrate transmission of data between segments by high-frequency carrier waves and, using a similar mechanism, we note that the passing of power between coupled coils in neighboring modules—or from an external power source—is also possible. Voice coils are a convenient multifunctional alternative to existing soft robot actuators. Their self-contained nature and ability to communicate with each other are ideal for modular robotics, and the additional functionality of sound input/output and power transfer will become increasingly useful as soft robots begin the transition from early proof-of-concept systems toward fully functional and highly integrated robotic systems. PMID:28078195

  17. Voice interactive electronic warning systems (VIEWS) - An applied approach to voice technology in the helicopter cockpit

    NASA Technical Reports Server (NTRS)

    Voorhees, J. W.; Bucher, N. M.

    1983-01-01

    The cockpit has been one of the most rapidly changing areas of new aircraft design over the past thirty years. In connection with these developments, a pilot can now be considered a decision maker/system manager as well as a vehicle controller. There is, however, a trend towards an information overload in the cockpit, and information processing problems begin to occur for the rotorcraft pilot. One approach to overcome the arising difficulties is based on the utilization of voice technology to improve the information transfer rate in the cockpit with respect to both input and output. Attention is given to the background of speech technology, the application of speech technology within the cockpit, voice interactive electronic warning system (VIEWS) simulation, and methodology. Information subsystems are considered along with a dynamic simulation study, and data collection.

  18. Optimization of multilayer neural network parameters for speaker recognition

    NASA Astrophysics Data System (ADS)

    Tovarek, Jaromir; Partila, Pavol; Rozhon, Jan; Voznak, Miroslav; Skapa, Jan; Uhrin, Dominik; Chmelikova, Zdenka

    2016-05-01

    This article discusses the impact of multilayer neural network parameters for speaker identification. The main task of speaker identification is to find a specific person in the known set of speakers. It means that the voice of an unknown speaker (wanted person) belongs to a group of reference speakers from the voice database. One of the requests was to develop the text-independent system, which means to classify wanted person regardless of content and language. Multilayer neural network has been used for speaker identification in this research. Artificial neural network (ANN) needs to set parameters like activation function of neurons, steepness of activation functions, learning rate, the maximum number of iterations and a number of neurons in the hidden and output layers. ANN accuracy and validation time are directly influenced by the parameter settings. Different roles require different settings. Identification accuracy and ANN validation time were evaluated with the same input data but different parameter settings. The goal was to find parameters for the neural network with the highest precision and shortest validation time. Input data of neural networks are a Mel-frequency cepstral coefficients (MFCC). These parameters describe the properties of the vocal tract. Audio samples were recorded for all speakers in a laboratory environment. Training, testing and validation data set were split into 70, 15 and 15 %. The result of the research described in this article is different parameter setting for the multilayer neural network for four speakers.

  19. Evaluation of a voice recognition system for the MOTAS pseudo pilot station function

    NASA Technical Reports Server (NTRS)

    Houck, J. A.

    1982-01-01

    The Langley Research Center has undertaken a technology development activity to provide a capability, the mission oriented terminal area simulation (MOTAS), wherein terminal area and aircraft systems studies can be performed. An experiment was conducted to evaluate state-of-the-art voice recognition technology and specifically, the Threshold 600 voice recognition system to serve as an aircraft control input device for the MOTAS pseudo pilot station function. The results of the experiment using ten subjects showed a recognition error of 3.67 percent for a 48-word vocabulary tested against a programmed vocabulary of 103 words. After the ten subjects retrained the Threshold 600 system for the words which were misrecognized or rejected, the recognition error decreased to 1.96 percent. The rejection rates for both cases were less than 0.70 percent. Based on the results of the experiment, voice recognition technology and specifically the Threshold 600 voice recognition system were chosen to fulfill this MOTAS function.

  20. Human factors issues associated with the use of speech technology in the cockpit

    NASA Technical Reports Server (NTRS)

    Kersteen, Z. A.; Damos, D.

    1983-01-01

    The human factors issues associated with the use of voice technology in the cockpit are summarized. The formulation of the LHX avionics suite is described and the allocation of tasks to voice in the cockpit is discussed. State-of-the-art speech recognition technology is reviewed. Finally, a questionnaire designed to tap pilot opinions concerning the allocation of tasks to voice input and output in the cockpit is presented. This questionnaire was designed to be administered to operational AH-1G Cobra gunship pilots. Half of the questionnaire deals specifically with the AH-1G cockpit and the types of tasks pilots would like to have performed by voice in this existing rotorcraft. The remaining portion of the questionnaire deals with an undefined rotorcraft of the future and is aimed at determining what types of tasks these pilots would like to have performed by voice technology if anything was possible, i.e. if there were no technological constraints.

  1. Using Continuous Voice Recognition Technology as an Input Medium to the Naval Warfare Interactive Simulation System (NWISS).

    DTIC Science & Technology

    1984-06-01

    Co ,u’arataor, Gr 7- / ’ . c ; / , caae.ic >ar. ’ ’# d:.i II ’ ..... .. . . .. .. . ... . , rV ABSTRACT A great d-al of research has been conducted an...9 2. Continuous Voice -%ecoait.ior, ....... 11 B. VERBEX 3000 SPEECH APPLiCATION DEVELOP !ENT SYSTEM! ( SPADS ...13 C . NAVAL IAR FARE INT7EACTI7E S:AIULATIC"N SYSTEM (NWISS) ....... .................. 14 D. PURPOSE .................... 16 1. A Past

  2. Voice input/output capabilities at Perception Technology Corporation

    NASA Technical Reports Server (NTRS)

    Ferber, Leon A.

    1977-01-01

    Condensed resumes of key company personnel at the Perception Technology Corporation are presented. The staff possesses recognition, speech synthesis, speaker authentication, and language identification. Hardware and software engineers' capabilities are included.

  3. An Investigation of the Application of Voice Input/Output Technology in the COINS Network Control Center,

    DTIC Science & Technology

    1982-03-01

    13: p. 27]. There are some connected-speech reccgnizers on the market today but they are expensive * 8 ($50,0-$10e,200) and their capabilities have...readout, end stock market quotationsrRef. 17: p. 6]. The second voice response technique, formant sjrthesis, uses a method in which a word library (again...users. Marketing brochures, therefore, should be looked 2t rather carefully, the best guarantee cf recogniticr. accuracy being a test with the desired

  4. Electronic data generation and display system

    NASA Technical Reports Server (NTRS)

    Wetekamm, Jules

    1988-01-01

    The Electronic Data Generation and Display System (EDGADS) is a field tested paperless technical manual system. The authoring provides subject matter experts the option of developing procedureware from digital or hardcopy inputs of technical information from text, graphics, pictures, and recorded media (video, audio, etc.). The display system provides multi-window presentations of graphics, pictures, animations, and action sequences with text and audio overlays on high resolution color CRT and monochrome portable displays. The database management system allows direct access via hierarchical menus, keyword name, ID number, voice command or touch of a screen pictoral of the item (ICON). It contains operations and maintenance technical information at three levels of intelligence for a total system.

  5. Voice loops as coordination aids in space shuttle mission control.

    PubMed

    Patterson, E S; Watts-Perotti, J; Woods, D D

    1999-01-01

    Voice loops, an auditory groupware technology, are essential coordination support tools for experienced practitioners in domains such as air traffic management, aircraft carrier operations and space shuttle mission control. They support synchronous communication on multiple channels among groups of people who are spatially distributed. In this paper, we suggest reasons for why the voice loop system is a successful medium for supporting coordination in space shuttle mission control based on over 130 hours of direct observation. Voice loops allow practitioners to listen in on relevant communications without disrupting their own activities or the activities of others. In addition, the voice loop system is structured around the mission control organization, and therefore directly supports the demands of the domain. By understanding how voice loops meet the particular demands of the mission control environment, insight can be gained for the design of groupware tools to support cooperative activity in other event-driven domains.

  6. Voice loops as coordination aids in space shuttle mission control

    NASA Technical Reports Server (NTRS)

    Patterson, E. S.; Watts-Perotti, J.; Woods, D. D.

    1999-01-01

    Voice loops, an auditory groupware technology, are essential coordination support tools for experienced practitioners in domains such as air traffic management, aircraft carrier operations and space shuttle mission control. They support synchronous communication on multiple channels among groups of people who are spatially distributed. In this paper, we suggest reasons for why the voice loop system is a successful medium for supporting coordination in space shuttle mission control based on over 130 hours of direct observation. Voice loops allow practitioners to listen in on relevant communications without disrupting their own activities or the activities of others. In addition, the voice loop system is structured around the mission control organization, and therefore directly supports the demands of the domain. By understanding how voice loops meet the particular demands of the mission control environment, insight can be gained for the design of groupware tools to support cooperative activity in other event-driven domains.

  7. Practical applications of interactive voice technologies: Some accomplishments and prospects

    NASA Technical Reports Server (NTRS)

    Grady, Michael W.; Hicklin, M. B.; Porter, J. E.

    1977-01-01

    A technology assessment of the application of computers and electronics to complex systems is presented. Three existing systems which utilize voice technology (speech recognition and speech generation) are described. Future directions in voice technology are also described.

  8. Silent reading of direct versus indirect speech activates voice-selective areas in the auditory cortex.

    PubMed

    Yao, Bo; Belin, Pascal; Scheepers, Christoph

    2011-10-01

    In human communication, direct speech (e.g., Mary said: "I'm hungry") is perceived to be more vivid than indirect speech (e.g., Mary said [that] she was hungry). However, for silent reading, the representational consequences of this distinction are still unclear. Although many of us share the intuition of an "inner voice," particularly during silent reading of direct speech statements in text, there has been little direct empirical confirmation of this experience so far. Combining fMRI with eye tracking in human volunteers, we show that silent reading of direct versus indirect speech engenders differential brain activation in voice-selective areas of the auditory cortex. This suggests that readers are indeed more likely to engage in perceptual simulations (or spontaneous imagery) of the reported speaker's voice when reading direct speech as opposed to meaning-equivalent indirect speech statements as part of a more vivid representation of the former. Our results may be interpreted in line with embodied cognition and form a starting point for more sophisticated interdisciplinary research on the nature of auditory mental simulation during reading.

  9. "Who" is saying "what"? Brain-based decoding of human voice and speech.

    PubMed

    Formisano, Elia; De Martino, Federico; Bonte, Milene; Goebel, Rainer

    2008-11-07

    Can we decipher speech content ("what" is being said) and speaker identity ("who" is saying it) from observations of brain activity of a listener? Here, we combine functional magnetic resonance imaging with a data-mining algorithm and retrieve what and whom a person is listening to from the neural fingerprints that speech and voice signals elicit in the listener's auditory cortex. These cortical fingerprints are spatially distributed and insensitive to acoustic variations of the input so as to permit the brain-based recognition of learned speech from unknown speakers and of learned voices from previously unheard utterances. Our findings unravel the detailed cortical layout and computational properties of the neural populations at the basis of human speech recognition and speaker identification.

  10. Voice Therapy: A Need for Research.

    ERIC Educational Resources Information Center

    Reed, Charles G.

    1980-01-01

    Conceptual and methodological guidelines for voice therapy research are presented, and suggestions are offered for selecting experimental designs. Divergent terminology, philosophy, and issues of voice therapy are examined to serve as an overview and as a basis for research direction. (Author/DLS)

  11. Guidelines for Selecting Microphones for Human Voice Production Research

    ERIC Educational Resources Information Center

    Svec, Jan G.; Granqvist, Svante

    2010-01-01

    Purpose: This tutorial addresses fundamental characteristics of microphones (frequency response, frequency range, dynamic range, and directionality), which are important for accurate measurements of voice and speech. Method: Technical and voice literature was reviewed and analyzed. The following recommendations on desirable microphone…

  12. Exploring interpersonal behavior and team sensemaking during health information technology implementation.

    PubMed

    Kitzmiller, Rebecca R; McDaniel, Reuben R; Johnson, Constance M; Lind, E Allan; Anderson, Ruth A

    2013-01-01

    We examine how interpersonal behavior and social interaction influence team sensemaking and subsequent team actions during a hospital-based health information technology (HIT) implementation project. Over the course of 18 months, we directly observed the interpersonal interactions of HIT implementation teams using a sensemaking lens. We identified three voice-promoting strategies enacted by team leaders that fostered team member voice and sensemaking; communicating a vision; connecting goals to team member values; and seeking team member input. However, infrequent leader expressions of anger quickly undermined team sensemaking, halting dialog essential to problem solving. By seeking team member opinions, team leaders overcame the negative effects of anger. Leaders must enact voice-promoting behaviors and use them throughout a team's engagement. Further, training teams in how to use conflict to achieve greater innovation may improve sensemaking essential to project risk mitigation. Health care work processes are complex; teams involved in implementing improvements must be prepared to deal with conflicting, contentious issues, which will arise during change. Therefore, team conflict training may be essential to sustaining sensemaking. Future research should seek to identify team interactions that foster sensemaking, especially when topics are difficult or unwelcome, then determine the association between staff sensemaking and the impact on HIT implementation outcomes. We are among the first to focus on project teams tasked with HIT implementation. This research extends our understanding of how leaders' behaviors might facilitate or impeded speaking up among project teams in health care settings.

  13. Implicit multisensory associations influence voice recognition.

    PubMed

    von Kriegstein, Katharina; Giraud, Anne-Lise

    2006-10-01

    Natural objects provide partially redundant information to the brain through different sensory modalities. For example, voices and faces both give information about the speech content, age, and gender of a person. Thanks to this redundancy, multimodal recognition is fast, robust, and automatic. In unimodal perception, however, only part of the information about an object is available. Here, we addressed whether, even under conditions of unimodal sensory input, crossmodal neural circuits that have been shaped by previous associative learning become activated and underpin a performance benefit. We measured brain activity with functional magnetic resonance imaging before, while, and after participants learned to associate either sensory redundant stimuli, i.e. voices and faces, or arbitrary multimodal combinations, i.e. voices and written names, ring tones, and cell phones or brand names of these cell phones. After learning, participants were better at recognizing unimodal auditory voices that had been paired with faces than those paired with written names, and association of voices with faces resulted in an increased functional coupling between voice and face areas. No such effects were observed for ring tones that had been paired with cell phones or names. These findings demonstrate that brief exposure to ecologically valid and sensory redundant stimulus pairs, such as voices and faces, induces specific multisensory associations. Consistent with predictive coding theories, associative representations become thereafter available for unimodal perception and facilitate object recognition. These data suggest that for natural objects effective predictive signals can be generated across sensory systems and proceed by optimization of functional connectivity between specialized cortical sensory modules.

  14. Research on oral test modeling based on multi-feature fusion

    NASA Astrophysics Data System (ADS)

    Shi, Yuliang; Tao, Yiyue; Lei, Jun

    2018-04-01

    In this paper, the spectrum of speech signal is taken as an input of feature extraction. The advantage of PCNN in image segmentation and other processing is used to process the speech spectrum and extract features. And a new method combining speech signal processing and image processing is explored. At the same time of using the features of the speech map, adding the MFCC to establish the spectral features and integrating them with the features of the spectrogram to further improve the accuracy of the spoken language recognition. Considering that the input features are more complicated and distinguishable, we use Support Vector Machine (SVM) to construct the classifier, and then compare the extracted test voice features with the standard voice features to achieve the spoken standard detection. Experiments show that the method of extracting features from spectrograms using PCNN is feasible, and the fusion of image features and spectral features can improve the detection accuracy.

  15. Proceedings of the Workshop on Meteorological and Environmental Inputs to Aviation Systems (6th) Held at the University of Tennessee Space Institute on 26-28 October 1982,

    DTIC Science & Technology

    1983-04-01

    it is mediocre information, as long 28 RESPOSE : Jim Sullivan data link can unload the voice channels. How- ever, this raises workload questions for...automa- This capability provides an alternative to tion program provides for pilot self -briefing voice communications. Where we’ve got crowded...etc., and various display-menu-product call-up 1983 use. and user-oriented self -help schemes. Additional PROFS products, including radar The Denver

  16. Delta modulation

    NASA Technical Reports Server (NTRS)

    Schilling, D. L.

    1971-01-01

    The conclusions of the design research of the song adaptive delta modulator are presented for source encoding voice signals. The variation of output SNR vs input signal power/when 8, 9, and 10 bit internal arithmetic is employed. Voice intelligibility tapes to test the 10-bit system are used. An analysis of a delta modulator is also presented designed to minimize the in-band rms error. This is accomplished by frequency shaping the error signal in the modulator prior to hard limiting. The result is a significant increase in the output SNR measured after low pass filtering.

  17. 14 CFR 29.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... second pilot stations and voice communications of other crewmembers on the flight deck when directed to those stations; or (2) By installing a continually energized or voice-actuated lip microphone at the first and second pilot stations. The microphone specified in this paragraph must be so located and, if...

  18. Implicit Multisensory Associations Influence Voice Recognition

    PubMed Central

    von Kriegstein, Katharina; Giraud, Anne-Lise

    2006-01-01

    Natural objects provide partially redundant information to the brain through different sensory modalities. For example, voices and faces both give information about the speech content, age, and gender of a person. Thanks to this redundancy, multimodal recognition is fast, robust, and automatic. In unimodal perception, however, only part of the information about an object is available. Here, we addressed whether, even under conditions of unimodal sensory input, crossmodal neural circuits that have been shaped by previous associative learning become activated and underpin a performance benefit. We measured brain activity with functional magnetic resonance imaging before, while, and after participants learned to associate either sensory redundant stimuli, i.e. voices and faces, or arbitrary multimodal combinations, i.e. voices and written names, ring tones, and cell phones or brand names of these cell phones. After learning, participants were better at recognizing unimodal auditory voices that had been paired with faces than those paired with written names, and association of voices with faces resulted in an increased functional coupling between voice and face areas. No such effects were observed for ring tones that had been paired with cell phones or names. These findings demonstrate that brief exposure to ecologically valid and sensory redundant stimulus pairs, such as voices and faces, induces specific multisensory associations. Consistent with predictive coding theories, associative representations become thereafter available for unimodal perception and facilitate object recognition. These data suggest that for natural objects effective predictive signals can be generated across sensory systems and proceed by optimization of functional connectivity between specialized cortical sensory modules. PMID:17002519

  19. [Signs and symptoms of autonomic dysfunction in dysphonic individuals].

    PubMed

    Park, Kelly; Behlau, Mara

    2011-01-01

    To verify the occurrence of signs and symptoms of autonomic nervous system dysfunction in individuals with behavioral dysphonia, and to compare it with the results obtained by individuals without vocal complaints. Participants were 128 adult individuals with ages between 14 and 74 years, divided into two groups: behavioral dysphonia (61 subjects) and without vocal complaints (67 subjects). It was administered the Protocol of Autonomic Dysfunction, containing 46 questions: 22 related to the autonomic nervous system and had no direct relationship with voice, 16 related to both autonomic nervous system and voice, six non-relevant questions, and two reliability questions. There was a higher occurrence of reported neurovegetative signs in the group with behavioral dysphonia, in questions related to voice, such as frequent throat clearing, frequent swallowing need, fatigability when speaking, and sore throat. In questions not directly related to voice, dysphonic individuals presented greater occurrence of three out of 22 symptoms: gas, tinnitus and aerophagia. Both groups presented similar results in questions non-relevant to the autonomic nervous system. Reliability questions needed reformulation. Individuals with behavioral dysphonia present higher occurrence of neurovegetative signs and symptoms, particularly those with direct relationship with voice, indicating greater lability of the autonomic nervous system in these subjects.

  20. Automation of Command and Data Entry in a Glovebox Work Volume: An Evaluation of Data Entry Devices

    NASA Technical Reports Server (NTRS)

    Steele, Marianne K.; Nakamura, Gail; Havens, Cindy; LeMay, Moira

    1996-01-01

    The present study was designed to examine the human-computer interface for data entry while performing experimental procedures within a glovebox work volume in order to make a recommendation to the Space Station Biological Research Project for a data entry system to be used within the Life Sciences Glovebox. Test subjects entered data using either a manual keypad, similar to a standard computer numerical keypad located within the glovebox work volume, or a voice input system using a speech recognition program with a microphone headset. Numerical input and commands were programmed in an identical manner between the two systems. With both electronic systems, a small trackball was available within the work volume for cursor control. Data, such as sample vial identification numbers, sample tissue weights, and health check parameters of the specimen, were entered directly into procedures that were electronically displayed on a video monitor within the glovebox. A pen and paper system with a 'flip-chart' format for procedure display, similar to that currently in use on the Space Shuttle, was used as a baseline data entry condition. Procedures were performed by a single operator; eight test subjects were used in the study. The electronic systems were tested under both a 'nominal' or 'anomalous' condition. The anomalous condition was introduced into the experimental procedure to increase the probability of finding limitations or problems with human interactions with the electronic systems. Each subject performed five test runs during a test day: two procedures each with voice and keypad, one with and one without anomalies, and one pen and paper procedure. The data collected were both quantitative (times, errors) and qualitative (subjective ratings of the subjects).

  1. Hidden Student Voice: A Curriculum of a Middle School Science Class Heard through Currere

    ERIC Educational Resources Information Center

    Crooks, Kathleen Schwartz

    2012-01-01

    Students have their own lenses through which they view school science and the students' views are often left out of educational conversations which directly affect the students themselves. Pinar's (2004) definition of curriculum as a "complicated conversation" implies that the class' voice is important, as important as the teacher's voice, to the…

  2. The effect of voice quality and competing speakers in a passage comprehension task: performance in relation to cognitive functioning in children with normal hearing.

    PubMed

    von Lochow, Heike; Lyberg-Åhlander, Viveka; Sahlén, Birgitta; Kastberg, Tobias; Brännström, K Jonas

    2018-04-01

    This study explores the effect of voice quality and competing speaker/-s on children's performance in a passage comprehension task. Furthermore, it explores the interaction between passage comprehension and cognitive functioning. Forty-nine children (27 girls and 22 boys) with normal hearing (aged 7-12 years) participated. Passage comprehension was tested in six different listening conditions; a typical voice (non-dysphonic voice) in quiet, a typical voice with one competing speaker, a typical voice with four competing speakers, a dysphonic voice in quiet, a dysphonic voice with one competing speaker, and a dysphonic voice with four competing speakers. The children's working memory capacity and executive functioning were also assessed. The findings indicate no direct effect of voice quality on the children's performance, but a significant effect of background listening condition. Interaction effects were seen between voice quality, background listening condition, and executive functioning. The children's susceptibility to the effect of the dysphonic voice and the background listening conditions are related to the individual's executive functions. The findings have several implications for design of interventions in language learning environments such as classrooms.

  3. Sperry Univac speech communications technology

    NASA Technical Reports Server (NTRS)

    Medress, Mark F.

    1977-01-01

    Technology and systems for effective verbal communication with computers were developed. A continuous speech recognition system for verbal input, a word spotting system to locate key words in conversational speech, prosodic tools to aid speech analysis, and a prerecorded voice response system for speech output are described.

  4. Voice - How humans communicate?

    PubMed

    Tiwari, Manjul; Tiwari, Maneesha

    2012-01-01

    Voices are important things for humans. They are the medium through which we do a lot of communicating with the outside world: our ideas, of course, and also our emotions and our personality. The voice is the very emblem of the speaker, indelibly woven into the fabric of speech. In this sense, each of our utterances of spoken language carries not only its own message but also, through accent, tone of voice and habitual voice quality it is at the same time an audible declaration of our membership of particular social regional groups, of our individual physical and psychological identity, and of our momentary mood. Voices are also one of the media through which we (successfully, most of the time) recognize other humans who are important to us-members of our family, media personalities, our friends, and enemies. Although evidence from DNA analysis is potentially vastly more eloquent in its power than evidence from voices, DNA cannot talk. It cannot be recorded planning, carrying out or confessing to a crime. It cannot be so apparently directly incriminating. As will quickly become evident, voices are extremely complex things, and some of the inherent limitations of the forensic-phonetic method are in part a consequence of the interaction between their complexity and the real world in which they are used. It is one of the aims of this article to explain how this comes about. This subject have unsolved questions, but there is no direct way to present the information that is necessary to understand how voices can be related, or not, to their owners.

  5. Ultrasonic speech translator and communications system

    DOEpatents

    Akerman, M.A.; Ayers, C.W.; Haynes, H.D.

    1996-07-23

    A wireless communication system undetectable by radio frequency methods for converting audio signals, including human voice, to electronic signals in the ultrasonic frequency range, transmitting the ultrasonic signal by way of acoustical pressure waves across a carrier medium, including gases, liquids, or solids, and reconverting the ultrasonic acoustical pressure waves back to the original audio signal. The ultrasonic speech translator and communication system includes an ultrasonic transmitting device and an ultrasonic receiving device. The ultrasonic transmitting device accepts as input an audio signal such as human voice input from a microphone or tape deck. The ultrasonic transmitting device frequency modulates an ultrasonic carrier signal with the audio signal producing a frequency modulated ultrasonic carrier signal, which is transmitted via acoustical pressure waves across a carrier medium such as gases, liquids or solids. The ultrasonic receiving device converts the frequency modulated ultrasonic acoustical pressure waves to a frequency modulated electronic signal, demodulates the audio signal from the ultrasonic carrier signal, and conditions the demodulated audio signal to reproduce the original audio signal at its output. 7 figs.

  6. Bilingual Voicing: A Study of Code-Switching in the Reported Speech of Finnish Immigrants in Estonia

    ERIC Educational Resources Information Center

    Frick, Maria; Riionheimo, Helka

    2013-01-01

    Through a conversation analytic investigation of Finnish-Estonian bilingual (direct) reported speech (i.e., voicing) by Finns who live in Estonia, this study shows how code-switching is used as a double contextualization device. The code-switched voicings are shaped by the on-going interactional situation, serving its needs by opening up a context…

  7. Design and fabrication of a new electrolarynx and voice amplifier for laryngectomees.

    PubMed

    Sundeep Krishna, M; Jayanthy, A K; Divakar, C; Mekhala, R

    2005-01-01

    A Laryngectomee is a person whose vocal cords i.e. voice box is surgically removed owing to cancer or due to automobile accidents, burns or trauma. The patient, therefore permanently loses the ability to speak normally. An Electrolarynx is an electronic speech aid that enables the Laryngectomee to communicate with other people as quickly as possible after the successful removal of the larynx. A neck type Electrolarynx has been designed. Earlier designs could not alter frequency and intensity simultaneously during conversation. The Electrolarynx developed can control both frequency and intensity simultaneously during conversation. The device has been tested on the patient and found to be very effective. A portable, pocket size, battery powered voice amplifier (PA system) has also been developed which uses an electret condenser microphone as the input. The voice amplifier developed is a two stage amplifier which uses a preamplifier stage and a power amplifier stage. The output of the power amplifier is connected to a speaker. The device is being used by the patient and found to be very useful.

  8. Vocal responses to unanticipated perturbations in voice loudness feedback: an automatic mechanism for stabilizing voice amplitude.

    PubMed

    Bauer, Jay J; Mittal, Jay; Larson, Charles R; Hain, Timothy C

    2006-04-01

    The present study tested whether subjects respond to unanticipated short perturbations in voice loudness feedback with compensatory responses in voice amplitude. The role of stimulus magnitude (+/- 1,3 vs 6 dB SPL), stimulus direction (up vs down), and the ongoing voice amplitude level (normal vs soft) were compared across compensations. Subjects responded to perturbations in voice loudness feedback with a compensatory change in voice amplitude 76% of the time. Mean latency of amplitude compensation was 157 ms. Mean response magnitudes were smallest for 1-dB stimulus perturbations (0.75 dB) and greatest for 6-dB conditions (0.98 dB). However, expressed as gain, responses for 1-dB perturbations were largest and almost approached 1.0. Response magnitudes were larger for the soft voice amplitude condition compared to the normal voice amplitude condition. A mathematical model of the audio-vocal system captured the main features of the compensations. Previous research has demonstrated that subjects can respond to an unanticipated perturbation in voice pitch feedback with an automatic compensatory response in voice fundamental frequency. Data from the present study suggest that voice loudness feedback can be used in a similar manner to monitor and stabilize voice amplitude around a desired loudness level.

  9. The use of wavelet packet transform and artificial neural networks in analysis and classification of dysphonic voices.

    PubMed

    Crovato, César David Paredes; Schuck, Adalberto

    2007-10-01

    This paper presents a dysphonic voice classification system using the wavelet packet transform and the best basis algorithm (BBA) as dimensionality reductor and 06 artificial neural networks (ANN) acting as specialist systems. Each ANN was a 03-layer multilayer perceptron with 64 input nodes, 01 output node and in the intermediary layer the number of neurons depends on the related training pathology group. The dysphonic voice database was separated in five pathology groups and one healthy control group. Each ANN was trained and associated with one of the 06 groups, and fed by the best base tree (BBT) nodes' entropy values, using the multiple cross validation (MCV) method and the leave-one-out (LOO) variation technique and success rates obtained were 87.5%, 95.31%, 87.5%, 100%, 96.87% and 89.06% for the groups 01 to 06, respectively.

  10. Lunar module voice recorder

    NASA Technical Reports Server (NTRS)

    1974-01-01

    A feasibility unit suitable for use as a voice recorder on the space shuttle was developed. A modification, development, and test program is described. A LM-DSEA recorder was modified to achieve the following goals: (1) redesign case to allow in-flight cartridge change; (2) time code change from LM code to IRIG-B 100 pps code; (3) delete cold plate requirements (also requires deletion of long-term thermal vacuum operation at 0.00001 MMHg); (4) implement track sequence reset during cartridge change; (5) reduce record time per cartridge because of unavailability of LM thin-base tape; and (6) add an internal Vox key circuit to turn on/off transport and electronics with voice data input signal. The recorder was tested at both the LM and shuttle vibration levels. The modified recorder achieved the same level of flutter during vibration as the DSEA recorder prior to modification. Several improvements were made over the specification requirements. The high manufacturing cost is discussed.

  11. High Tech and Library Access for People with Disabilities.

    ERIC Educational Resources Information Center

    Roatch, Mary A.

    1992-01-01

    Describes tools that enable people with disabilities to access print information, including optical character recognition, synthetic voice output, other input devices, Braille access devices, large print displays, television and video, TDD (Telecommunications Devices for the Deaf), and Telebraille. Use of technology by libraries to meet mandates…

  12. Community Coauthoring: Whose Voice Remains?

    ERIC Educational Resources Information Center

    Larson, Joanne; Webster, Stephanie; Hopper, Mindy

    2011-01-01

    This article examines how texts are collaboratively produced in community development work when coauthors come from multiple racial, ethnic, and class backgrounds as well as business and other work experiences. We found that the term "wordsmithing" became a discursive tool that limited resident input and shaped the Plan toward an…

  13. Voice emotion perception and production in cochlear implant users.

    PubMed

    Jiam, N T; Caldwell, M; Deroche, M L; Chatterjee, M; Limb, C J

    2017-09-01

    Voice emotion is a fundamental component of human social interaction and social development. Unfortunately, cochlear implant users are often forced to interface with highly degraded prosodic cues as a result of device constraints in extraction, processing, and transmission. As such, individuals with cochlear implants frequently demonstrate significant difficulty in recognizing voice emotions in comparison to their normal hearing counterparts. Cochlear implant-mediated perception and production of voice emotion is an important but relatively understudied area of research. However, a rich understanding of the voice emotion auditory processing offers opportunities to improve upon CI biomedical design and to develop training programs benefiting CI performance. In this review, we will address the issues, current literature, and future directions for improved voice emotion processing in cochlear implant users. Copyright © 2017 Elsevier B.V. All rights reserved.

  14. One positive impact of health care reform to physicians: the computer-based patient record.

    PubMed

    England, S P

    1993-11-01

    The health care industry is an information-dependent business that will require a new generation of health information systems if successful health care reform is to occur. We critically need integrated clinical management information systems to support the physician and related clinicians at the direct care level, which in turn will have linkages with secondary users of health information such as health payors, regulators, and researchers. The economic dependence of health care industry on the CPR cannot be underestimated, says Jeffrey Ritter. He sees the U.S. health industry as about to enter a bold new age where our records are electronic, our computers are interconnected, and our money is nothing but pulses running across the telephone lines. Hence the United States is now in an age of electronic commerce. Clinical systems reform must begin with the community-based patient chart, which is located in the physician's office, the hospital, and other related health care provider offices. A community-based CPR and CPR system that integrates all providers within a managed care network is the most logical step since all health information begins with the creation of a patient record. Once a community-based CPR system is in place, the physician and his or her clinical associates will have a common patient record upon which all direct providers have access to input and record patient information. Once a community-level CPR system is in place with a community provider network, each physician will have available health information and data processing capability that will finally provide real savings in professional time and effort. Lost patient charts will no longer be a problem. Data input and storage of health information would occur electronically via transcripted text, voice, and document imaging. All electronic clinical information, voice, and graphics could be recalled at any time and transmitted to any terminal location within the health provider network. Hence, health system re-engineering must begin and be developed where health information is initially created--in the physician's office or clinic.

  15. High or low? Comparing high and low-variability phonetic training in adult and child second language learners.

    PubMed

    Giannakopoulou, Anastasia; Brown, Helen; Clayards, Meghan; Wonnacott, Elizabeth

    2017-01-01

    High talker variability (i.e., multiple voices in the input) has been found effective in training nonnative phonetic contrasts in adults. A small number of studies suggest that children also benefit from high-variability phonetic training with some evidence that they show greater learning (more plasticity) than adults given matched input, although results are mixed. However, no study has directly compared the effectiveness of high versus low talker variability in children. Native Greek-speaking eight-year-olds ( N = 52), and adults ( N = 41) were exposed to the English /i/-/ɪ/ contrast in 10 training sessions through a computerized word-learning game. Pre- and post-training tests examined discrimination of the contrast as well as lexical learning. Participants were randomly assigned to high (four talkers) or low (one talker) variability training conditions. Both age groups improved during training, and both improved more while trained with a single talker. Results of a three-interval oddity discrimination test did not show the predicted benefit of high-variability training in either age group. Instead, children showed an effect in the reverse direction-i.e., reliably greater improvements in discrimination following single talker training, even for untrained generalization items, although the result is qualified by (accidental) differences between participant groups at pre-test. Adults showed a numeric advantage for high-variability but were inconsistent with respect to voice and word novelty. In addition, no effect of variability was found for lexical learning. There was no evidence of greater plasticity for phonetic learning in child learners. This paper adds to the handful of studies demonstrating that, like adults, child learners can improve their discrimination of a phonetic contrast via computerized training. There was no evidence of a benefit of training with multiple talkers, either for discrimination or word learning. The results also do not support the findings of greater plasticity in child learners found in a previous paper (Giannakopoulou, Uther & Ylinen, 2013a). We discuss these results in terms of various differences between training and test tasks used in the current work compared with previous literature.

  16. Mindfulness of voices, self-compassion, and secure attachment in relation to the experience of hearing voices.

    PubMed

    Dudley, James; Eames, Catrin; Mulligan, John; Fisher, Naomi

    2018-03-01

    Developing compassion towards oneself has been linked to improvement in many areas of psychological well-being, including psychosis. Furthermore, developing a non-judgemental, accepting way of relating to voices is associated with lower levels of distress for people who hear voices. These factors have also been associated with secure attachment. This study explores associations between the constructs of mindfulness of voices, self-compassion, and distress from hearing voices and how secure attachment style related to each of these variables. Cross-sectional online. One hundred and twenty-eight people (73% female; M age  = 37.5; 87.5% Caucasian) who currently hear voices completed the Self-Compassion Scale, Southampton Mindfulness of Voices Questionnaire, Relationships Questionnaire, and Hamilton Programme for Schizophrenia Voices Questionnaire. Results showed that mindfulness of voices mediated the relationship between self-compassion and severity of voices, and self-compassion mediated the relationship between mindfulness of voices and severity of voices. Self-compassion and mindfulness of voices were significantly positively correlated with each other and negatively correlated with distress and severity of voices. Mindful relation to voices and self-compassion are associated with reduced distress and severity of voices, which supports the proposed potential benefits of mindful relating to voices and self-compassion as therapeutic skills for people experiencing distress by voice hearing. Greater self-compassion and mindfulness of voices were significantly associated with less distress from voices. These findings support theory underlining compassionate mind training. Mindfulness of voices mediated the relationship between self-compassion and distress from voices, indicating a synergistic relationship between the constructs. Although the current findings do not give a direction of causation, consideration is given to the potential impact of mindful and compassionate approaches to voices. © 2017 The Authors. British Journal of Clinical Psychology published by John Wiley & Sons Ltd on behalf of British Psychological Society.

  17. On shame and voice-hearing

    PubMed Central

    2017-01-01

    Hearing voices in the absence of another speaker—what psychiatry terms an auditory verbal hallucination—is often associated with a wide range of negative emotions. Mainstream clinical research addressing the emotional dimensions of voice-hearing has tended to treat these as self-evident, undifferentiated and so effectively interchangeable. But what happens when a richer, more nuanced understanding of specific emotions is brought to bear on the analysis of distressing voices? This article draws findings from the ‘What is it like to hear voices’ study conducted as part of the interdisciplinary Hearing the Voice project into conversation with philosopher Dan Zahavi's Self and Other: Exploring Subjectivity, Empathy and Shame to consider how a focus on shame can open up new questions about the experience of hearing voices. A higher-order emotion of social cognition, shame directs our attention to aspects of voice-hearing which are understudied and elusive, particularly as they concern the status of voices as other and the constitution and conceptualisation of the self. PMID:28389551

  18. Validation of Mobility of Pedestrians with Low Vision Using Graphic Floor Signs and Voice Guides.

    PubMed

    Omori, Kiyohiro; Yanagihara, Takao; Kitagawa, Hiroshi; Ikeda, Norihiro

    2015-01-01

    Some people with low vision or elderly persons tend to walk while watching a nearby floor, therefore, they often overlook or hard to read suspended signs. In this study, we propose two kinds of voice guides, and an experiment is conducted by participants with low vision using these voice guides and graphic floor signs in order to investigate effectiveness of these combinations. In clock position method (CP), each direction of near facilities are described in using an analogy of a 12-hour clock. Meanwhile, in numbering method (NU), near facilities are put the number in clockwise order, however, each direction are only illustrated in a crossing sign. As a result of an experiment, it is showed that both voice guides are effective for pedestrians with low vision. NU is used as a complement of graphic floor signs. Meanwhile, CP is used independently with graphic floor signs, however, there is a risk in the case of using in the environment where pedestrians are easy to mistake the reference direction defined by the sounding speaker.

  19. Concentrating on Affective Feedforward in Online Tutoring

    ERIC Educational Resources Information Center

    Chen, Ya-Ting; Chou, Yung-Hsin; Cowan, John

    2014-01-01

    With considerable input from the student voice, the paper centres on a detailed account of the experiences of Western academic, tutoring Eastern students online to develop their critical thinking skills. From their online experiences together as tutor and students, the writers present a considered case for the main emphasis in facilitative online…

  20. Transforming Belief Systems in Minneapolis

    ERIC Educational Resources Information Center

    Walker, Michael; Yeager, Corey; Zumbusch, Jennie

    2018-01-01

    The Office of Black Male Student Achievement (OBMSA) of Minneapolis Public Schools (MPS), established in 2014, is one of the first in the country. The innovative work of the OBMSA is centered on student voice and student thought. After getting input from parents and families, community members, educators, and young Black males themselves, the…

  1. ACCC's Response to Industry Canada's Consultation on Improving Canada's Digital Advantage

    ERIC Educational Resources Information Center

    Association of Canadian Community Colleges, 2010

    2010-01-01

    As the national and international voice representing over 150 publicly-funded colleges, institutes, polytechnics, cegeps, university colleges and universities with a college mandate, the Association of Canadian Community Colleges (ACCC) welcomes the opportunity to provide input to Industry Canada's consultation on a Digital Economy Strategy for…

  2. 78 FR 12271 - Wireline Competition Bureau Seeks Additional Comment In Connect America Cost Model Virtual Workshop

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-02-22

    ... Competition Bureau seeks public input on additional questions relating to modeling voice capability and Annual... submitting comments and additional information on the rulemaking process, see the SUPPLEMENTARY INFORMATION section of this document. FOR FURTHER INFORMATION CONTACT: Katie King, Wireline Competition Bureau at (202...

  3. Twenty-Channel Voice Response System

    DOT National Transportation Integrated Search

    1981-06-01

    This report documents the design and implementation of a Voice Response System, which provides Direct-User Access to the FAA's aviation-weather data base. This system supports 20 independent audio channels, and as of this report, speaks three weather...

  4. Acoustic and Perceived Measurements Certifying Tango as Voice Treatment Method.

    PubMed

    Tafiadis, Dionysios; Kosma, Evangelia I; Chronopoulos, Spyridon K; Papadopoulos, Aggelos; Toki, Eugenia I; Vassiliki, Siafaka; Ziavra, Nausica

    2018-03-01

    Voice disorders are affecting everyday life in many levels, and their prevalence has been studied extensively in certain and general populations. Notably, several factors have a cohesive influence on voice disorders and voice characteristics. Several studies report that health and environmental and psychological etiologies can serve as risk factors for voice disorders. Many diagnostic protocols, in the literature, evaluate voice and its parameters leading to direct or indirect treatment intervention. This study was designed to examine the effect of tango on adult acoustic voice parameters. Fifty-two adults (26 male and 26 female) were recruited and divided into four subgroups (male dancers, female dancers, male nondancers, and female nondancers). The participants were asked to answer two questionnaires (Voice Handicap Index and Voice Evaluation Form), and their voices were recorded before and after the tango dance session. Moreover, water consumption was investigated. The study's results indicated that the voices' acoustic characteristics were different between tango dancers and the control group. The beneficial results are far from prominent as they prove that tango dance can serve stand-alone as voice therapy without the need for hydration. Also, more research is imperative to be conducted on a longitudinal basis to obtain a more accurate result on the required time for the proposed therapy. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  5. Bringing voice in policy building.

    PubMed

    Lotrecchiano, Gaetano R; Kane, Mary; Zocchi, Mark S; Gosa, Jessica; Lazar, Danielle; Pines, Jesse M

    2017-07-03

    Purpose The purpose of this paper is to describe the use of group concept mapping (GCM) as a tool for developing a conceptual model of an episode of acute, unscheduled care from illness or injury to outcomes such as recovery, death and chronic illness. Design/methodology/approach After generating a literature review drafting an initial conceptual model, GCM software (CS Global MAX TM ) is used to organize and identify strengths and directionality between concepts generated through feedback about the model from several stakeholder groups: acute care and non-acute care providers, patients, payers and policymakers. Through online and in-person population-specific focus groups, the GCM approach seeks feedback, assigned relationships and articulated priorities from participants to produce an output map that described overarching concepts and relationships within and across subsamples. Findings A clustered concept map made up of relational data points that produced a taxonomy of feedback was used to update the model for use in soliciting additional feedback from two technical expert panels (TEPs), and finally, a public comment exercise was performed. The results were a stakeholder-informed improved model for an acute care episode, identified factors that influence process and outcomes, and policy recommendations, which were delivered to the Department of Health and Human Services's (DHHS) Assistant Secretary for Preparedness and Response. Practical implications This study provides an example of the value of cross-population multi-stakeholder input to increase voice in shared problem health stakeholder groups. Originality/value This paper provides GCM results and a visual analysis of the relational characteristics both within and across sub-populations involved in the study. It also provides an assessment of observational key factors supporting how different stakeholder voices can be integrated to inform model development and policy recommendations.

  6. Vocal indices of stress: a review.

    PubMed

    Giddens, Cheryl L; Barron, Kirk W; Byrd-Craven, Jennifer; Clark, Keith F; Winter, A Scott

    2013-05-01

    Identification of stress patterns in the voice has multiple potential applications. The objective was to review literature pertaining to the effects of various forms of stress upon the healthy voice. Literature review, discussion of results, and direction for further study. This review article offers a model of stress and a review of the historical and recent research into the effects of stress on the voice. Electronic databases were searched using the key words. No studies were excluded on the basis of design; however, an attempt was made to include in the discussion studies which primarily address physiological and acoustic vocal parameters. The results of greater than 50 studies examining the effect of stressors ranging from lie and guilt to high altitude and space flight upon the voice were included in the review. Increase in fundamental frequency is the most commonly reported effect of stress in well-controlled trials. The trend, however, is not universal. A reduction in noise as reflected by the diminished vocal jitter is reported, but less frequently. Stress types, gender, and individual differences in baseline autonomic tone may explain the primarily equivocal findings of effects of stressor exposure or perceived stress on voice; and as such, the article concludes with a discussion of directions for future study. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  7. Maximal Ambient Noise Levels and Type of Voice Material Required for Valid Use of Smartphones in Clinical Voice Research.

    PubMed

    Lebacq, Jean; Schoentgen, Jean; Cantarella, Giovanna; Bruss, Franz Thomas; Manfredi, Claudia; DeJonckere, Philippe

    2017-09-01

    Smartphone technology provides new opportunities for recording standardized voice samples of patients and transmitting the audio files to the voice laboratory. This drastically improves the achievement of baseline designs, used in research on efficiency of voice treatments. However, the basic requirement is the suitability of smartphones for recording and digitizing pathologic voices (mainly characterized by period perturbations and noise) without significant distortion. In a previous article, this was tested using realistic synthesized deviant voice samples (/a:/) with three precisely known levels of jitter and of noise in all combinations. High correlations were found between jitter and noise to harmonics ratio measured in (1) recordings via smartphones, (2) direct microphone recordings, and (3) sound files generated by the synthesizer. In the present work, similar experiments were performed (1) in the presence of increasing levels of ambient noise and (2) using synthetic deviant voice samples (/a:/) as well as synthetic voice material simulating a deviant short voiced utterance (/aiuaiuaiu/). Ambient noise levels up to 50 dB A are acceptable. However, signal processing occurs in some smartphones, and this significantly affects estimates of jitter and noise to harmonics ratio when formant changes are introduced in analogy with running speech. The conclusion is that voice material must provisionally be limited to a sustained /a/. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  8. Assessments of Voice Use and Voice Quality among College/University Singing Students Ages 18–24 through Ambulatory Monitoring with a Full Accelerometer Signal

    PubMed Central

    Schloneger, Matthew; Hunter, Eric

    2016-01-01

    The multiple social and performance demands placed on college/university singers could put their still developing voices at risk. Previous ambulatory monitoring studies have analyzed the duration, intensity, and frequency (in Hz) of voice use among such students. Nevertheless, no studies to date have incorporated the simultaneous acoustic voice quality measures into the acquisition of these measures to allow for direct comparison during the same voicing period. Such data could provide greater insight into how young singers use their voices, as well as identify potential correlations between vocal dose and acoustic changes in voice quality. The purpose of this study was to assess the voice use and estimated voice quality of college/university singing students (18–24 y/o, N = 19). Ambulatory monitoring was conducted over three full, consecutive weekdays measuring voice from an unprocessed accelerometer signal measured at the neck. From this signal were analyzed traditional vocal dose metrics such as phonation percentage, dose time, cycle dose, and distance dose. Additional acoustic measures included perceived pitch, pitch strength, LTAS slope, alpha ratio, dB SPL 1–3 kHz, and harmonic-to-noise ratio. Major findings from more than 800 hours of recording indicated that among these students (a) higher vocal doses correlated significantly with greater voice intensity, more vocal clarity and less perturbation; and (b) there were significant differences in some acoustic voice quality metrics between non-singing, solo singing and choral singing. PMID:26897545

  9. Development and Testing of a Portable Vocal Accumulator

    ERIC Educational Resources Information Center

    Cheyne, Harold A.; Hanson, Helen M.; Genereux, Ronald P.; Stevens, Kenneth N.; Hillman, Robert E.

    2003-01-01

    This research note describes the design and testing of a device for unobtrusive, long-term ambulatory monitoring of voice use, named the Portable Vocal Accumulator (PVA). The PVA contains a digital signal processor for analyzing input from a neck-placed miniature accelerometer. During its development, accelerometer recordings were obtained from 99…

  10. Information Processing Concepts: A Cure for "Technofright." Information Processing in the Electronic Office. Part 1: Concepts.

    ERIC Educational Resources Information Center

    Popyk, Marilyn K.

    1986-01-01

    Discusses the new automated office and its six major technologies (data processing, word processing, graphics, image, voice, and networking), the information processing cycle (input, processing, output, distribution/communication, and storage and retrieval), ergonomics, and ways to expand office education classes (versus class instruction). (CT)

  11. Rethinking Roles, Relationships and Voices in Studies of Undergraduate Student Writers

    ERIC Educational Resources Information Center

    Looker, Samantha

    2012-01-01

    Undergraduate students have a complex and often problematic history of representation in research on writing pedagogy. They have been described as novices and outsiders, while having minimal input into how they are studied and represented. In this piece, I share my efforts to rethink the roles and relationships among researchers and student…

  12. Listening to Student Voices: How Student Advisory Boards Can Help.

    ERIC Educational Resources Information Center

    Bacon, Ellen; Bloom, Lisa

    2000-01-01

    This article describes the involvement of students with emotional and/or behavior disorders on effective student advisory boards. Examples are given of student advisory board input in elementary school conflict mediation and mentor programs, a middle school composure room program, and a high school in-school factory program. Stressed is the…

  13. Language and Communication-Related Problems of Aviation Safety.

    ERIC Educational Resources Information Center

    Cushing, Steven

    A study of the problems posed by the use of natural language in various aspects of aviation is presented. The study, part of a larger investigation of the feasibility of voice input/output interfaces for communication in aviation, looks at representative real examples of accidents and near misses resulting from language confusions and omissions.…

  14. Listen to Your Inner Voice: Using Your Intuition in Outdoor Leadership.

    ERIC Educational Resources Information Center

    Cook, Janice

    Intuition is knowledge of something without the conscious use of reasoning. The question of where intuitive knowledge comes from may be addressed from neurophysiological, spiritual, or philosophical perspectives. In some cases, hunches may be traced to the unconscious processing of immediate sensory input with previous knowledge. In other cases,…

  15. Learning with Portable Digital Devices in Australian Schools: 20 Years On!

    ERIC Educational Resources Information Center

    Newhouse, C. Paul

    2014-01-01

    Portable computing technologies such as laptops, tablets, smartphones, wireless networking, voice/stylus input, and plug and play peripheral devices, appear to offer the means of finally realising much of the long heralded vision for computers to support learning in schools. There is the possibility for the technology to finally become a…

  16. Ultrasonic speech translator and communications system

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Akerman, M.A.; Ayers, C.W.; Haynes, H.D.

    1996-07-23

    A wireless communication system undetectable by radio frequency methods for converting audio signals, including human voice, to electronic signals in the ultrasonic frequency range, transmitting the ultrasonic signal by way of acoustical pressure waves across a carrier medium, including gases, liquids, or solids, and reconverting the ultrasonic acoustical pressure waves back to the original audio signal. The ultrasonic speech translator and communication system includes an ultrasonic transmitting device and an ultrasonic receiving device. The ultrasonic transmitting device accepts as input an audio signal such as human voice input from a microphone or tape deck. The ultrasonic transmitting device frequency modulatesmore » an ultrasonic carrier signal with the audio signal producing a frequency modulated ultrasonic carrier signal, which is transmitted via acoustical pressure waves across a carrier medium such as gases, liquids or solids. The ultrasonic receiving device converts the frequency modulated ultrasonic acoustical pressure waves to a frequency modulated electronic signal, demodulates the audio signal from the ultrasonic carrier signal, and conditions the demodulated audio signal to reproduce the original audio signal at its output. 7 figs.« less

  17. Ultrasonic speech translator and communications system

    DOEpatents

    Akerman, M. Alfred; Ayers, Curtis W.; Haynes, Howard D.

    1996-01-01

    A wireless communication system undetectable by radio frequency methods for converting audio signals, including human voice, to electronic signals in the ultrasonic frequency range, transmitting the ultrasonic signal by way of acoustical pressure waves across a carrier medium, including gases, liquids, or solids, and reconverting the ultrasonic acoustical pressure waves back to the original audio signal. The ultrasonic speech translator and communication system (20) includes an ultrasonic transmitting device (100) and an ultrasonic receiving device (200). The ultrasonic transmitting device (100) accepts as input (115) an audio signal such as human voice input from a microphone (114) or tape deck. The ultrasonic transmitting device (100) frequency modulates an ultrasonic carrier signal with the audio signal producing a frequency modulated ultrasonic carrier signal, which is transmitted via acoustical pressure waves across a carrier medium such as gases, liquids or solids. The ultrasonic receiving device (200) converts the frequency modulated ultrasonic acoustical pressure waves to a frequency modulated electronic signal, demodulates the audio signal from the ultrasonic carrier signal, and conditions the demodulated audio signal to reproduce the original audio signal at its output (250).

  18. High-speed digital phonoscopy images analyzed by Nyquist plots

    NASA Astrophysics Data System (ADS)

    Yan, Yuling

    2012-02-01

    Vocal-fold vibration is a key dynamic event in voice production, and the vibratory characteristics of the vocal fold correlate closely with voice quality and health condition. Laryngeal imaging provides direct means to observe the vocal fold vibration; in the past, however, available modalities were either too slow or impractical to resolve the actual vocal fold vibrations. This limitation has now been overcome by high-speed digital imaging (HSDI) (or high-speed digital phonoscopy), which records images of the vibrating vocal folds at a rate of 2000 frames per second or higher- fast enough to resolve a specific, sustained phonatory vocal fold vibration. The subsequent image-based functional analysis of voice is essential to better understanding the mechanism underlying voice production, as well as assisting the clinical diagnosis of voice disorders. Our primary objective is to develop a comprehensive analytical platform for voice analysis using the HSDI recordings. So far, we have developed various analytical approaches for the HSDI-based voice analyses. These include Nyquist plots and associated analysese that are used along with FFT and Spectrogram in the analysis of the HSDI data representing normal voice and specific voice pathologies.

  19. Assessments of Voice Use and Voice Quality Among College/University Singing Students Ages 18-24 Through Ambulatory Monitoring With a Full Accelerometer Signal.

    PubMed

    Schloneger, Matthew J; Hunter, Eric J

    2017-01-01

    The multiple social and performance demands placed on college/university singers could put their still-developing voices at risk. Previous ambulatory monitoring studies have analyzed the duration, intensity, and frequency (in Hertz) of voice use among such students. Nevertheless, no studies to date have incorporated the simultaneous acoustic voice quality measures into the acquisition of these measures to allow for direct comparison during the same voicing period. Such data could provide greater insight into how young singers use their voices, as well as identify potential correlations between vocal dose and acoustic changes in voice quality. The purpose of this study was to assess the voice use and the estimated voice quality of college/university singing students (18-24 years old, N = 19). Ambulatory monitoring was conducted over three full, consecutive weekdays measuring voice from an unprocessed accelerometer signal measured at the neck. From this signal, traditional vocal dose metrics such as phonation percentage, dose time, cycle dose, and distance dose were analyzed. Additional acoustic measures included perceived pitch, pitch strength, long-term average spectrum slope, alpha ratio, dB sound pressure level 1-3 kHz, and harmonic-to-noise ratio. Major findings from more than 800 hours of recording indicated that among these students (a) higher vocal doses correlated significantly with greater voice intensity, more vocal clarity and less perturbation; and (b) there were significant differences in some acoustic voice quality metrics between nonsinging, solo singing, and choral singing. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  20. Andreas Vesalius' 500th Anniversary: Initial Integral Understanding of Voice Production.

    PubMed

    Brinkman, Romy J; Hage, J Joris

    2017-01-01

    Voice production relies on the integrated functioning of a three-part system: respiration, phonation and resonance, and articulation. To commemorate the 500th anniversary of the great anatomist Andreas Vesalius (1515-1564), we report on his understanding of this integral system. The text of Vesalius' masterpiece De Humani Corporis Fabrica Libri Septum and an eyewitness report of the public dissection of three corpses by Vesalius in Bologna, Italy, in 1540, were searched for references to the voice-producing anatomical structures and their function. We clustered the traced, separate parts for the first time. We found that Vesalius recognized the importance for voice production of many details of the respiratory system, the voice box, and various structures of resonance and articulation. He stressed that voice production was a cerebral function and extensively recorded the innervation of the voice-producing organs by the cranial nerves. Vesalius was the first to publicly record the concept of voice production as an integrated and cerebrally directed function of respiration, phonation and resonance, and articulation. In doing so nearly 500 years ago, he laid a firm basis for the understanding of the physiology of voice production and speech and its management as we know it today. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  1. Visual face-movement sensitive cortex is relevant for auditory-only speech recognition.

    PubMed

    Riedel, Philipp; Ragert, Patrick; Schelinski, Stefanie; Kiebel, Stefan J; von Kriegstein, Katharina

    2015-07-01

    It is commonly assumed that the recruitment of visual areas during audition is not relevant for performing auditory tasks ('auditory-only view'). According to an alternative view, however, the recruitment of visual cortices is thought to optimize auditory-only task performance ('auditory-visual view'). This alternative view is based on functional magnetic resonance imaging (fMRI) studies. These studies have shown, for example, that even if there is only auditory input available, face-movement sensitive areas within the posterior superior temporal sulcus (pSTS) are involved in understanding what is said (auditory-only speech recognition). This is particularly the case when speakers are known audio-visually, that is, after brief voice-face learning. Here we tested whether the left pSTS involvement is causally related to performance in auditory-only speech recognition when speakers are known by face. To test this hypothesis, we applied cathodal transcranial direct current stimulation (tDCS) to the pSTS during (i) visual-only speech recognition of a speaker known only visually to participants and (ii) auditory-only speech recognition of speakers they learned by voice and face. We defined the cathode as active electrode to down-regulate cortical excitability by hyperpolarization of neurons. tDCS to the pSTS interfered with visual-only speech recognition performance compared to a control group without pSTS stimulation (tDCS to BA6/44 or sham). Critically, compared to controls, pSTS stimulation additionally decreased auditory-only speech recognition performance selectively for voice-face learned speakers. These results are important in two ways. First, they provide direct evidence that the pSTS is causally involved in visual-only speech recognition; this confirms a long-standing prediction of current face-processing models. Secondly, they show that visual face-sensitive pSTS is causally involved in optimizing auditory-only speech recognition. These results are in line with the 'auditory-visual view' of auditory speech perception, which assumes that auditory speech recognition is optimized by using predictions from previously encoded speaker-specific audio-visual internal models. Copyright © 2015 Elsevier Ltd. All rights reserved.

  2. Automatic detection of voice impairments by means of short-term cepstral parameters and neural network based detectors.

    PubMed

    Godino-Llorente, J I; Gómez-Vilda, P

    2004-02-01

    It is well known that vocal and voice diseases do not necessarily cause perceptible changes in the acoustic voice signal. Acoustic analysis is a useful tool to diagnose voice diseases being a complementary technique to other methods based on direct observation of the vocal folds by laryngoscopy. Through the present paper two neural-network based classification approaches applied to the automatic detection of voice disorders will be studied. Structures studied are multilayer perceptron and learning vector quantization fed using short-term vectors calculated accordingly to the well-known Mel Frequency Coefficient cepstral parameterization. The paper shows that these architectures allow the detection of voice disorders--including glottic cancer--under highly reliable conditions. Within this context, the Learning Vector quantization methodology demonstrated to be more reliable than the multilayer perceptron architecture yielding 96% frame accuracy under similar working conditions.

  3. The superiority in voice processing of the blind arises from neural plasticity at sensory processing stages.

    PubMed

    Föcker, Julia; Best, Anna; Hölig, Cordula; Röder, Brigitte

    2012-07-01

    Blind people rely much more on voices compared to sighted individuals when identifying other people. Previous research has suggested a faster processing of auditory input in blind individuals than sighted controls and an enhanced activation of temporal cortical regions during voice processing. The present study used event-related potentials (ERPs) to single out the sub-processes of auditory person identification that change and allow for superior voice processing after congenital blindness. A priming paradigm was employed in which two successive voices (S1 and S2) of either the same (50% of the trials) or different actors were presented. Congenitally blind and matched sighted participants made an old-young decision on the S2. During the pre-experimental familiarization with the stimuli, congenitally blind individuals showed faster learning rates than sighted controls. Reaction times were shorter in person-congruent trials than in person-incongruent trials in both groups. ERPs to S2 stimuli in person-incongruent as compared to person-congruent trials were significantly enhanced at early processing stages (100-160 ms) in congenitally blind participants only. A later negative ERP effect (>200 ms) was found in both groups. The scalp topographies of the experimental effects were characterized by a central and parietal distribution in the sighted but a more posterior distribution in the congenitally blind. These results provide evidence for an improvement of early voice processing stages and a reorganization of the person identification system as a neural correlate of compensatory behavioral improvements following congenital blindness. Copyright © 2012 Elsevier Ltd. All rights reserved.

  4. Speaking more broadly: an examination of the nature, antecedents, and consequences of an expanded set of employee voice behaviors.

    PubMed

    Maynes, Timothy D; Podsakoff, Philip M

    2014-01-01

    Scholarly interest in employee voice behavior has increased dramatically over the past 15 years. Although this research has produced valuable knowledge, it has focused almost exclusively on voice as a positively intended challenge to the status quo, even though some scholars have argued that it need not challenge the status quo or be well intentioned. Thus, in this paper, we create an expanded view of voice; one that extends beyond voice as a positively intended challenge to the status quo to include voice that supports how things are being done in organizations as well as voice that may not be well intentioned. We construct a framework based on this expanded view that identifies 4 different types of voice behavior (supportive, constructive, defensive, and destructive). We then develop and validate survey measures for each of these. Evidence from 5 studies across 4 samples provides strong support for our new measures in that (a) a 4-factor confirmatory factor analysis model fit the data significantly better than 1-, 2-, or 3-factor models; (b) the voice measures converged with and yet remained distinct from conceptually related comparison constructs; (c) personality predictors exhibited unique patterns of relationships with the different types of voice; (d) variations in actual voice behaviors had a direct causal impact on responses to the survey items; and (e) each type of voice significantly impacted important outcomes for voicing employees (e.g., likelihood of relying on a voicing employee's opinions and evaluations of a voicing employee's overall performance). Implications of our findings are discussed. PsycINFO Database Record (c) 2014 APA, all rights reserved

  5. Using voice input and audio feedback to enhance the reality of a virtual experience

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Miner, N.E.

    1994-04-01

    Virtual Reality (VR) is a rapidly emerging technology which allows participants to experience a virtual environment through stimulation of the participant`s senses. Intuitive and natural interactions with the virtual world help to create a realistic experience. Typically, a participant is immersed in a virtual environment through the use of a 3-D viewer. Realistic, computer-generated environment models and accurate tracking of a participant`s view are important factors for adding realism to a virtual experience. Stimulating a participant`s sense of sound and providing a natural form of communication for interacting with the virtual world are equally important. This paper discusses the advantagesmore » and importance of incorporating voice recognition and audio feedback capabilities into a virtual world experience. Various approaches and levels of complexity are discussed. Examples of the use of voice and sound are presented through the description of a research application developed in the VR laboratory at Sandia National Laboratories.« less

  6. Low Vocal Pitch Preference Drives First Impressions Irrespective of Context in Male Voices but Not in Female Voices.

    PubMed

    Tsantani, Maria S; Belin, Pascal; Paterson, Helena M; McAleer, Phil

    2016-08-01

    Vocal pitch has been found to influence judgments of perceived trustworthiness and dominance from a novel voice. However, the majority of findings arise from using only male voices and in context-specific scenarios. In two experiments, we first explore the influence of average vocal pitch on first-impression judgments of perceived trustworthiness and dominance, before establishing the existence of an overall preference for high or low pitch across genders. In Experiment 1, pairs of high- and low-pitched temporally reversed recordings of male and female vocal utterances were presented in a two-alternative forced-choice task. Results revealed a tendency to select the low-pitched voice over the high-pitched voice as more trustworthy, for both genders, and more dominant, for male voices only. Experiment 2 tested an overall preference for low-pitched voices, and whether judgments were modulated by speech content, using forward and reversed speech to manipulate context. Results revealed an overall preference for low pitch, irrespective of direction of speech, in male voices only. No such overall preference was found for female voices. We propose that an overall preference for low pitch is a default prior in male voices irrespective of context, whereas pitch preferences in female voices are more context- and situation-dependent. The present study confirms the important role of vocal pitch in the formation of first-impression personality judgments and advances understanding of the impact of context on pitch preferences across genders.

  7. Speaking Up: How Patient and Physician Voices Shaped a Trial to Improve Goals-of-Care Discussions.

    PubMed

    Solomon, Rachel; Smith, Cardinale; Kallio, Jay; Fenollosa, Amy; Benerofe, Barbara; Jones, Laurence; Adelson, Kerin; Gonsky, Jason P; Messner, Carolyn; Bickell, Nina A

    2017-08-01

    Patients with advanced cancer benefit from early goals-of-care (GoC) conversations, but few facilitators are known. We describe the process and outcomes of involving patient and physician stakeholders in the design and development of a trial, funded by the Patient-Centered Outcomes Research Institute (PCORI), to enhance oncologists' communication skills and their propensity to facilitate productive, meaningful GoC discussions with patients with advanced cancer. We recruited oncologists, palliative care physicians, and patient stakeholders to participate in proposal development, intervention design and modification, identification of outcome measures, and refinement of study tools. Formats for exchange included 1:1 structured interviews, workshops, and stakeholder meetings. Patient and physician voices helped craft and implement a study of an intervention to enhance oncologists' ability to facilitate GoC discussions with patients with advanced cancer. Physician inputs guided the creation of an oncologist and palliative care physician "joint visit" intervention at a turning point in disease management. Patient inputs impacted on the language used, outcome measures assessed, and approaches used to introduce patients to the intervention visit. Stakeholder input informed the development of a novel intervention that physicians seemed to find both valuable and in sync with their needs and their practice schedules. Where communication about difficult subjects and shared decision making are involved, including multiple stakeholder groups in study design, implementation, and outcomes measurement may have far-reaching effects.

  8. Development and Validation of the Children's Voice Handicap Index-10 for Parents.

    PubMed

    Ricci-Maccarini, Andrea; De Maio, Vincenzo; Murry, Thomas; Schindler, Antonio

    2016-01-01

    The Children's Voice Handicap Index-10 (CVHI-10) was introduced as a tool for self-assessment of children's dysphonia. However, in the management of children with voice disorders, both parents' and children's perspectives play an important role. Because a self-tool including both a children's and a parents' version does not exist yet, the aim of the study was to develop and validate an assessment tool which parallels the CVHI-10 for parents to assess the level of voice handicap in their child's voice. Observational, prospective, cross-sectional study. To develop a CVHI-10 for parents, called "CVHI-10-P", the CVHI-10 items were adapted to reflect parents' responses about their child. Fifty-five children aged 7-12 years completed the CVHI-10, whereas their parents completed the CVHI-10-P. Each child's voice was also perceptually assessed by an otolaryngologist using the Grade Breathness Roughness (GRB) scale. Fifty-one of the 55 children underwent voice therapy (VT) and were assessed afterward using the GRB, CVHI-10, and CVHI-10-P. CVHI-10-P internal consistency was satisfactory (Cronbach alpha = .78). Correlation between CVHI-10-P and CVHI-10 was moderate (r = 0.37). CVHI-10-P total scores were lower than CVHI-10 scores in most of the cases. Single-item mean scores were always lower in CVHI-10-P compared with CVHI-10, with the exception of the only one item of the CVHI-10-P that directly involves the parent's experience (item 10). Data gained from one tool are not directly related to the other, suggesting that these two tools appraise the child's voice handicap from different perspectives. The overall perceptual assessment scores of the 51 children after VT significantly improved. There was a statistically significant reduction of the total scores and for each item in CVHI-10 and CVHI-10-P after VT. These data support the adoption of the CVHI-10-P as an assessment tool and an outcome measure for management of children's voice disorders. CVHI-10-P is a valid tool to appraise parents' perspective of their child's voice disorder. The use of the CVHI-10 and the CVHI-10-P is recommended for objectively determining the level of voice handicap in children by parents and child. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  9. Comparison of long-term voice outcomes after vocal fold augmentation using autologous fat injection by direct microlaryngoscopy versus office-based calcium hydroxylapatite injection.

    PubMed

    Zeleník, Karol; Walderová, Radana; Kučová, Hana; Jančatová, Debora; Komínek, Pavel

    2017-08-01

    The objective is to compare the long-term voice outcomes of vocal fold augmentation (VFA) using autologous fat injection via direct microlaryngoscopy versus office-based calcium hydroxylapatite (CaHA) injection. Patients with glottal insufficiency and a gap no greater than 3 mm caused by unilateral vocal fold paralysis or vocal fold atrophy were prospectively recruited to the study from September 2012 to September 2015. From September 2012 to May 2014, VFA was only performed using autologous fat via direct microlaryngoscopy under general anesthesia (N = 14). From May 2014 to September 2015, VFA was performed as an office-based procedure using a transoral approach to inject CaHA (N = 17). Videolaryngostroboscopic evaluation, subjective satisfaction with voice, voice handicap index (VHI), and maximal phonation time (MPT) were analyzed pre-injection and 12 months after VFA. A total of 31 patients were analyzed. One year after VFA, 67.8% of the patients were satisfied with their voice, with no significant difference between groups (P = 0.247). The mean improvement in VHI in the autologous fat group was 31.6 ± 16.82 versus 35 ± 27.24 in the CaHA group (P = 0.664). MPT improvement was also similar in the two groups: 5.5 ± 2.52 for the autologous fat group versus 6.0 ± 3.98 for the CaHA group (P = 0.823). Both autologous fat injection via direct microlaryngoscopy and office-based CaHA injection have good long-term results. There were no differences in the treatment results of the two procedures 1 year after injection.

  10. 14 CFR 27.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... stations and voice communications of other crewmembers on the flight deck when directed to those stations... pilot stations. The microphone specified in this paragraph must be so located and, if necessary, the... are intelligible when recorded under flight cockpit noise conditions and played back. The level of...

  11. 14 CFR 25.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... stations and voice communications of other crewmembers on the flight deck when directed to those stations... as practicable when recorded under flight cockpit noise conditions and played back. Repeated aural or... pilot station. (2) For the second channel from each boom, mask, or hand-held microphone, headset, or...

  12. Multimodal interfaces with voice and gesture input

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Milota, A.D.; Blattner, M.M.

    1995-07-20

    The modalities of speech and gesture have different strengths and weaknesses, but combined they create synergy where each modality corrects the weaknesses of the other. We believe that a multimodal system such a one interwining speech and gesture must start from a different foundation than ones which are based solely on pen input. In order to provide a basis for the design of a speech and gesture system, we have examined the research in other disciplines such as anthropology and linguistics. The result of this investigation was a taxonomy that gave us material for the incorporation of gestures whose meaningsmore » are largely transparent to the users. This study describes the taxonomy and gives examples of applications to pen input systems.« less

  13. Stakeholders' Voices: Defining Needs of Students with Emotional and Behavioral Disorders Transitioning between School Settings

    ERIC Educational Resources Information Center

    Buchanan, Rohanna; Nese, Rhonda N. T.; Clark, Miriam

    2016-01-01

    Students with emotional and behavioral disorders (EBD) too often do not receive adequate services or care in their school settings, particularly during transitions in educational placements. In addition, school support teams often struggle with creating transition plans that honor the needs of students with input from key stakeholders responsible…

  14. A Research Program in Computer Technology. Volume 1

    DTIC Science & Technology

    1981-08-01

    rigidity, sensor networks 10. command and control, digital voice communication, graphic input device for terminal, multimedia communications, portable...satellite channel in the internetwork environment; Distributed Sensor Networks - formulation of algorithms and communication protocols to support the...operation of geographically distributed sensors ; Personal Communicator - work intended to result in a demonstration-level portable terminal to test and

  15. An Adult Education Study of Participatory Community Mapping for Indigenous Knowledge Production

    ERIC Educational Resources Information Center

    Campbell, Craig A., Jr.

    2010-01-01

    This dissertation explores the notion of participatory community mapping (PCM) for Indigenous knowledge production. Three major questions were posed in the study. First, how can PCM foster Indigenous knowledge production and documentation? Second, how can PCM be used to include local voice and input in mapping projects, and third, how can adult…

  16. The COMMAND trial of cognitive therapy to prevent harmful compliance with command hallucinations: predictors of outcome and mediators of change.

    PubMed

    Birchwood, Max; Dunn, Graham; Meaden, Alan; Tarrier, Nicholas; Lewis, Shon; Wykes, Til; Davies, Linda; Michail, Maria; Peters, Emmanuelle

    2017-12-05

    Acting on harmful command hallucinations is a major clinical concern. Our COMMAND CBT trial approximately halved the rate of harmful compliance (OR = 0.45, 95% CI 0.23-0.88, p = 0.021). The focus of the therapy was a single mechanism, the power dimension of voice appraisal, was also significantly reduced. We hypothesised that voice power differential (between voice and voice hearer) was the mediator of the treatment effect. The trial sample (n = 197) was used. A logistic regression model predicting 18-month compliance was used to identify predictors, and an exploratory principal component analysis (PCA) of baseline variables used as potential predictors (confounders) in their own right. Stata's paramed command used to obtain estimates of the direct, indirect and total effects of treatment. Voice omnipotence was the best predictor although the PCA identified a highly predictive cognitive-affective dimension comprising: voices' power, childhood trauma, depression and self-harm. In the mediation analysis, the indirect effect of treatment was fully explained by its effect on the hypothesised mediator: voice power differential. Voice power and treatment allocation were the best predictors of harmful compliance up to 18 months; post-treatment, voice power differential measured at nine months was the mediator of the effect of treatment on compliance at 18 months.

  17. Music Signal Processing Using Vector Product Neural Networks

    NASA Astrophysics Data System (ADS)

    Fan, Z. C.; Chan, T. S.; Yang, Y. H.; Jang, J. S. R.

    2017-05-01

    We propose a novel neural network model for music signal processing using vector product neurons and dimensionality transformations. Here, the inputs are first mapped from real values into three-dimensional vectors then fed into a three-dimensional vector product neural network where the inputs, outputs, and weights are all three-dimensional values. Next, the final outputs are mapped back to the reals. Two methods for dimensionality transformation are proposed, one via context windows and the other via spectral coloring. Experimental results on the iKala dataset for blind singing voice separation confirm the efficacy of our model.

  18. In defense of the passive voice in medical writing.

    PubMed

    Minton, Timothy D

    2015-01-01

    Few medical journals specifically instruct authors to use the active voice and avoid the passive voice, but advice to that effect is common in the large number of stylebooks and blogs aimed at medical and scientific writers. Such advice typically revolves around arguments that the passive voice is less clear, less direct, and less concise than the active voice, that it conceals the identity of the person(s) performing the action(s) described, that it obscures meaning, that it is pompous, and that the high rate of passive-voice usage in scientific writing is a result of conformity to an established and old-fashioned style of writing. Some of these arguments are valid with respect to specific examples of passive-voice misuse by some medical (and other) writers, but as arguments for avoiding passive-voice use in general, they are seriously flawed. In addition, many of the examples that stylebook writers give of inappropriate use are actually much more appropriate in certain contexts than the active-voice alternatives they provide. In this review, I examine the advice offered by anti-passive writers, along with some of their examples of "inappropriate" use, and argue that the key factor in voice selection is sentence word order as determined by the natural tendency in English for the topic of discourse ("old" information) to take subject position and for "new" information to come later. Authors who submit to this natural tendency will not have to worry much about voice selection, because it will usually be automatic.

  19. 14 CFR 23.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... originating at the first and second pilot stations and voice communications of other crewmembers on the flight deck when directed to those stations. The microphone must be so located and, if necessary, the... conditions and played back. Repeated aural or visual playback of the record may be used in evaluating...

  20. 76 FR 18490 - Contributions to the Telecommunications Relay Service Fund

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-04-04

    ... voice over Internet Protocol (VoIP) service provider and each provider of non- interconnected VoIP... directs that within one year after the date of enactment of the CVAA, such VoIP providers shall... Fund (TRS Fund) by non-interconnected Voice over Internet Protocol (VoIP) service providers with...

  1. Evidence for Prosody in Silent Reading

    ERIC Educational Resources Information Center

    Gross, Jennifer; Millett, Amanda L.; Bartek, Brian; Bredell, Kyle Hampton; Winegard, Bo

    2014-01-01

    English speakers and expressive readers emphasize new content in an ongoing discourse. Do silent readers emphasize new content in their inner voice? Because the inner voice cannot be directly observed, we borrowed the cap-emphasis technique (e.g., "toMAYto") from the pronunciation guides of dictionaries to elicit prosodic emphasis.…

  2. Mechanical and dynamic aspects of voice production as related to voice therapy and phonosurgery.

    PubMed

    Isshiki, N

    2000-06-01

    Laryngeal framework surgery can change the position and tension of the vocal folds safely without direct surgical intervention in the vocal fold proper. Some 23 years of experience with phonosurgery have proved its usefulness in treating dysphonia related to unilateral vocal fold paralysis, vocal fold atrophy, and pitch-related dysphonias. Meanwhile, much information about the mechanism of voice production has been obtained through intraoperative findings of voice and fiberscopic examination of the larynx. Based on such knowledge together with information obtained through model experiments, the human vocal organ was reconsidered mainly from the mechanical view point, and the roles of voice therapy and singing pedagogy were discussed in relation to phonosurgery. The vocal organ may not be an ideal musical organ and is rather vulnerable, but its potential is enormous.

  3. Mechanical and dynamic aspects of voice production as related to voice therapy and phonosurgery.

    PubMed

    Isshiki, N

    1998-06-01

    Laryngeal framework surgery can change the position and tension of the vocal folds safely without direct surgical intervention in the vocal fold proper. Some 23 years of experience with phonosurgery have proved its usefulness in treating dysphonia related to unilateral vocal fold paralysis, vocal fold atrophy, and pitch-related dysphonias . Meanwhile, much information about the mechanism of voice production has been obtained through intraoperative findings of voice and fiberscopic examination of the larynx . Based on such knowledge together with information obtained through model experiments, the human vocal organ was reconsidered mainly from the mechanical view point, and the roles of voice therapy and singing pedagogy were discussed in relation to phonosurgery. The vocal organ may not be an ideal musical organ and is rather vulnerable, but its potential is enormous.

  4. Using Natural Language to Enable Mission Managers to Control Multiple Heterogeneous UAVs

    NASA Technical Reports Server (NTRS)

    Trujillo, Anna C.; Puig-Navarro, Javier; Mehdi, S. Bilal; Mcquarry, A. Kyle

    2016-01-01

    The availability of highly capable, yet relatively cheap, unmanned aerial vehicles (UAVs) is opening up new areas of use for hobbyists and for commercial activities. This research is developing methods beyond classical control-stick pilot inputs, to allow operators to manage complex missions without in-depth vehicle expertise. These missions may entail several heterogeneous UAVs flying coordinated patterns or flying multiple trajectories deconflicted in time or space to predefined locations. This paper describes the functionality and preliminary usability measures of an interface that allows an operator to define a mission using speech inputs. With a defined and simple vocabulary, operators can input the vast majority of mission parameters using simple, intuitive voice commands. Although the operator interface is simple, it is based upon autonomous algorithms that allow the mission to proceed with minimal input from the operator. This paper also describes these underlying algorithms that allow an operator to manage several UAVs.

  5. Electroglottographic analysis of actresses and nonactresses' voices in different levels of intensity.

    PubMed

    Master, Suely; Guzman, Marco; Carlos de Miranda, Helder; Lloyd, Adam

    2013-03-01

    Previous studies with long-term average spectrum (LTAS) showed the importance of the glottal source for understanding the projected voices of actresses. In this study, electroglottographic (EGG) analysis was used to investigate the contribution of the glottal source to the projected voice, comparing actresses and nonactresses' voices, in different levels of intensity. Thirty actresses and 30 nonactresses sustained vowels in habitual, moderate, and loud intensity levels. The EGG variables were contact quotient (CQ), closing quotient (QCQ), and opening quotient (QOQ). Other variables were sound pressure level (SPL) and fundamental frequency (F0). A KayPENTAX EGG was used. Variables were inputted in a general linear model. Actresses showed significantly higher values for SPL, in all levels, and both groups increased SPL significantly while changing from habitual to moderate and further to loud. There were no significant differences between groups for EGG quotients. There were significant differences between the levels only for F0 and CQ for both groups. SPL was significantly higher among actresses in all intensity levels, but in the EGG analysis, no differences were found. This apparently weak contribution of the glottal source in the supposedly projected voices of actresses, contrary to previous LTAS studies, might be because of a higher subglottal pressure or perhaps greater vocal tract contribution in SPL. Results from the present study suggest that trained subjects did not produce a significant higher SPL than untrained individuals by increasing the cost in terms of higher vocal fold collision and hence more impact stress. Future researches should explore the difference between trained and nontrained voices by aerodynamic measurements to evaluate the relationship between physiologic findings and the acoustic and EGG data. Moreover, further studies should consider both types of vocal tasks, sustained vowel and running speech, for both EGG and LTAS analysis. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  6. Start/End Delays of Voiced and Unvoiced Speech Signals

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Herrnstein, A

    Recent experiments using low power EM-radar like sensors (e.g, GEMs) have demonstrated a new method for measuring vocal fold activity and the onset times of voiced speech, as vocal fold contact begins to take place. Similarly the end time of a voiced speech segment can be measured. Secondly it appears that in most normal uses of American English speech, unvoiced-speech segments directly precede or directly follow voiced-speech segments. For many applications, it is useful to know typical duration times of these unvoiced speech segments. A corpus, assembled earlier of spoken ''Timit'' words, phrases, and sentences and recorded using simultaneously measuredmore » acoustic and EM-sensor glottal signals, from 16 male speakers, was used for this study. By inspecting the onset (or end) of unvoiced speech, using the acoustic signal, and the onset (or end) of voiced speech using the EM sensor signal, the average duration times for unvoiced segments preceding onset of vocalization were found to be 300ms, and for following segments, 500ms. An unvoiced speech period is then defined in time, first by using the onset of the EM-sensed glottal signal, as the onset-time marker for the voiced speech segment and end marker for the unvoiced segment. Then, by subtracting 300ms from the onset time mark of voicing, the unvoiced speech segment start time is found. Similarly, the times for a following unvoiced speech segment can be found. While data of this nature have proven to be useful for work in our laboratory, a great deal of additional work remains to validate such data for use with general populations of users. These procedures have been useful for applying optimal processing algorithms over time segments of unvoiced, voiced, and non-speech acoustic signals. For example, these data appear to be of use in speaker validation, in vocoding, and in denoising algorithms.« less

  7. Evaluation of Direct and Indirect Methods of Sub-Neoglottic Pressure Measurement in Tracheoesophageal Speakers: A Systematic Review and Meta-Analysis.

    PubMed

    Sheela, Shekaraiah; Aithal, Venkataraja U; Rajashekhar, Bellur; Lewis, Melissa Glenda

    2016-01-01

    Tracheoesophageal (TE) prosthetic voice is one of the voice restoration options for individuals who have undergone a total laryngectomy. Aerodynamic analysis of the TE voice provides insight into the physiological changes that occur at the level of the neoglottis with voice prosthesis in situ. The present study is a systematic review and meta-analysis of sub-neoglottic pressure (SNP) measurement in TE speakers by direct and indirect methods. The screening of abstracts and titles was carried out for inclusion of articles using 10 electronic databases spanning the period from 1979 to 2016. Ten articles which met the inclusion criteria were considered for meta-analysis with a pooled age range of 40-83 years. The pooled mean SNP obtained from the direct measurement method was 53.80 cm H2O with a 95% confidence interval of 21.14-86.46 cm H2O, while for the indirect measurement method, the mean SNP was 23.55 cm H2O with a 95% confidence interval of 19.23-27.87 cm H2O. Based on the literature review, the various procedures followed for direct and indirect measurements of SNP contributed to a range of differences in outcome measures. The meta-analysis revealed that the "interpolation method" for indirect estimation of SNP was the most acceptable and valid method in TE speakers. © 2017 S. Karger AG, Basel.

  8. Awakening Teacher Voice and Student Voice: The Development of a Feminist Pedagogy

    ERIC Educational Resources Information Center

    Weisner, Jill

    2004-01-01

    In this paper, the author presents a narrative describing both her personal life and professional life, sharing the development of her teaching strategies and their relationship to feminism. She details how her background, politics, insecurities, strengths, and values directly influenced the development of her teaching pedagogy, asserting that,…

  9. Student Voice and the Politics of Listening in Higher Education

    ERIC Educational Resources Information Center

    McLeod, Julie

    2011-01-01

    The promise of giving voice to under-represented and marginalized groups has been a mainstay of emancipatory agendas in educational research. It has been an especially influential focus in feminist and gender equity reform projects and is increasingly a feature of policies and programs directed to enhance youth participation and civic inclusion.…

  10. In Their Words: Student Choice in Training Markets--Victorian Examples. NCVER Research Report

    ERIC Educational Resources Information Center

    Brown, Justin

    2017-01-01

    This research offers insights into the options available to individuals as they navigate the vocational education and training (VET) market. Importantly, this study directly represents the voice of students, asking how their choices were made and whether their choice was sufficiently "informed." The student voice is contrasted with…

  11. Behavioral Treatment of Voice Disorders in Teachers

    PubMed Central

    Ziegler, Aaron; Gillespie, Amanda I.; Verdolini Abbott, Katherine

    2010-01-01

    Introduction The purpose of this paper is to review the literature on the behavioral treatment of voice disorders in teachers. The focus is on phonogenic disorders, that is voice disorders thought to be caused by voice use. Methods Review of the literature and commentary. Results The review exposes distinct holes in the literature on the treatment of voice problems in teachers. However, emerging trends in treatment are noted. For example, most studies identified for review implemented a multiple-therapy approach in a group setting, in contrast to only a few studies that assessed a single-therapy approach with individual patients. Although the review reveals that the evidence around behavioral treatment of voice disorders in teachers is mixed, a growing body of data provides some indicators on how effectively rehabilitation of teachers with phonogenic voice problems might be approached. Specifically, voice amplification demonstrates promise as a beneficial type of indirect therapy and vocal function exercises as well as resonant voice therapy show possible benefits as direct therapies. Finally, only a few studies identified even remotely begin to meet guidelines of the Consolidated Standards of Reporting Trials statement, a finding that emphasizes the need to increase the number of investigations that adhere to strict research standards. Conclusions Although data on the treatment of voice problems in teachers are still limited in the literature, emerging trends are noted. The accumulation of sufficient studies will ultimately provide useful evidence about this societally important issue. PMID:20093840

  12. Next generation keyboards: The importance of cognitive compatibility

    NASA Technical Reports Server (NTRS)

    Amell, John R.; Ewry, Michael E.; Colle, Herbert A.

    1988-01-01

    The computer keyboard of today is essentially the same as it has been for many years. Few advances have been made in keyboard design even though computer systems in general have made remarkable progress in improvements. This paper discusses the future of keyboards, their competition and compatibility with voice input systems, and possible special-application intelligent keyboards for controlling complex systems.

  13. Input and Output Mechanisms and Devices. Phase I: Adding Voice Output to a Speaker-Independent Recognition System.

    ERIC Educational Resources Information Center

    Scott Instruments Corp., Denton, TX.

    This project was designed to develop techniques for adding low-cost speech synthesis to educational software. Four tasks were identified for the study: (1) select a microcomputer with a built-in analog-to-digital converter that is currently being used in educational environments; (2) determine the feasibility of implementing expansion and playback…

  14. Investing in What It Takes to Move from Good to Great: Exemplary Educators Identify Their Most Important Learning Experiences

    ERIC Educational Resources Information Center

    Jacques, Catherine; Behrstock-Sherratt, Ellen; Parker, Amber; Bassett, Katherine; Allen, Megan; Bosso, David; Olson, Derek

    2017-01-01

    For the last 4 years, 10 leading education organizations have collaborated on a study series that includes teacher voice in conversations and research about educator effectiveness. Initially conceptualized by teacher leaders from the National Network of State Teachers of the Year (NNSTOY) and with their continued input, the "From Good to…

  15. Dynamic Pressure Microphones

    NASA Astrophysics Data System (ADS)

    Werner, E.

    In 1876, Alexander Graham Bell described his first telephone with a microphone using magnetic induction to convert the voice input into an electric output signal. The basic principle led to a variety of designs optimized for different needs, from hearing impaired users to singers or broadcast announcers. From the various sound pressure versions, only the moving coil design is still in mass production for speech and music application.

  16. Dysphonia, Perceived Control, and Psychosocial Distress: A Qualitative Study.

    PubMed

    Misono, Stephanie; Haut, Caroline; Meredith, Liza; Frazier, Patricia A; Stockness, Ali; Michael, Deirdre D; Butcher, Lisa; Harwood, Eileen M

    2018-05-11

    The purpose of this qualitative study was to examine relationships between psychological factors, particularly perceived control, and voice symptoms in adults seeking treatment for a voice problem. Semistructured interviews of adult patients with a clinical diagnosis of muscle tension dysphonia were conducted and transcribed. Follow-up interviews were conducted as needed for further information or clarification. A multidisciplinary team analyzed interview content using inductive techniques. Common themes and subthemes were identified. A conceptual model was developed describing the association between voice symptoms, psychological factors, precipitants of ongoing voice symptoms, and perceived control. Thematic saturation was reached after 23 interviews. No participants reported a direct psychological cause for their voice problem, although half described significant life events preceding voice problem onset (eg, miscarriage and other health events, interpersonal conflicts, and family members' illnesses, injuries, and deaths). Participants described psychological influences on voice symptoms that led to rapid exacerbation of their voice symptoms. Participants described the helpfulness of speech therapy and sometimes also challenges of applying techniques in daily life. They also discussed personal coping strategies that included behavioral (eg, avoiding triggers and seeking social support) and psychological (eg, mind-body awareness and emotion regulation) components. Voice-related perceived control was associated with adaptive emotional and behavioral responses, which appeared to facilitate symptom improvement. In this qualitative pilot study, participant narratives suggested that psychological factors and emotions influence voice symptoms, facilitating development of a preliminary conceptual model of how adaptive and maladaptive responses develop and how they influence vocal function. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  17. The need for experience focused counselling (EFC) with voice hearers in training and practice: a review of the literature.

    PubMed

    Schnackenberg, J K; Martin, C R

    2014-06-01

    A pathologizing paradigm to making sense of experiences such as hearing voices and schizophrenia remains dominant within mental health service provision. However, a real biological basis to the aetiology of hearing voices, and similar phenomena remains elusive. Antipsychotic medication, as the mainstay of the biological model, has not only been shown to have serious side effects, but is widely acknowledged as being of clinical benefit only to a limited number of people. In contrast, the Recovery Movement, and in particular the Hearing Voices Movement, have suggested that a normal life is possible despite having the experience of hearing voices. At its heart is the notion that it is possible to make sense of voices within the person's life context and to learn to live with them. Interestingly, it would seem that this approach remains largely confined to the user movement. This may in part be the result of the lack of widely accepted quantifiable and qualitative research in this area supporting such a stance. This review focuses on the current evidence base for the individual approach of the Hearing Voices Movement, which is known as Experience Focused Counselling or Making Sense of Voices. Future directions for research are indicated. © 2013 John Wiley & Sons Ltd.

  18. Adaptations in humans for assessing physical strength from the voice

    PubMed Central

    Sell, Aaron; Bryant, Gregory A.; Cosmides, Leda; Tooby, John; Sznycer, Daniel; von Rueden, Christopher; Krauss, Andre; Gurven, Michael

    2010-01-01

    Recent research has shown that humans, like many other animals, have a specialization for assessing fighting ability from visual cues. Because it is probable that the voice contains cues of strength and formidability that are not available visually, we predicted that selection has also equipped humans with the ability to estimate physical strength from the voice. We found that subjects accurately assessed upper-body strength in voices taken from eight samples across four distinct populations and language groups: the Tsimane of Bolivia, Andean herder-horticulturalists and United States and Romanian college students. Regardless of whether raters were told to assess height, weight, strength or fighting ability, they produced similar ratings that tracked upper-body strength independent of height and weight. Male voices were more accurately assessed than female voices, which is consistent with ethnographic data showing a greater tendency among males to engage in violent aggression. Raters extracted information about strength from the voice that was not supplied from visual cues, and were accurate with both familiar and unfamiliar languages. These results provide, to our knowledge, the first direct evidence that both men and women can accurately assess men's physical strength from the voice, and suggest that estimates of strength are used to assess fighting ability. PMID:20554544

  19. More than Just Two Sexes: The Neural Correlates of Voice Gender Perception in Gender Dysphoria

    PubMed Central

    Junger, Jessica; Habel, Ute; Bröhr, Sabine; Neulen, Josef; Neuschaefer-Rube, Christiane; Birkholz, Peter; Kohler, Christian; Schneider, Frank; Derntl, Birgit; Pauly, Katharina

    2014-01-01

    Gender dysphoria (also known as “transsexualism”) is characterized as a discrepancy between anatomical sex and gender identity. Research points towards neurobiological influences. Due to the sexually dimorphic characteristics of the human voice, voice gender perception provides a biologically relevant function, e.g. in the context of mating selection. There is evidence for a better recognition of voices of the opposite sex and a differentiation of the sexes in its underlying functional cerebral correlates, namely the prefrontal and middle temporal areas. This fMRI study investigated the neural correlates of voice gender perception in 32 male-to-female gender dysphoric individuals (MtFs) compared to 20 non-gender dysphoric men and 19 non-gender dysphoric women. Participants indicated the sex of 240 voice stimuli modified in semitone steps in the direction to the other gender. Compared to men and women, MtFs showed differences in a neural network including the medial prefrontal gyrus, the insula, and the precuneus when responding to male vs. female voices. With increased voice morphing men recruited more prefrontal areas compared to women and MtFs, while MtFs revealed a pattern more similar to women. On a behavioral and neuronal level, our results support the feeling of MtFs reporting they cannot identify with their assigned sex. PMID:25375171

  20. The Vocal Tract Organ: A New Musical Instrument Using 3-D Printed Vocal Tracts.

    PubMed

    Howard, David M

    2017-10-27

    The advent and now increasingly widespread availability of 3-D printers is transforming our understanding of the natural world by enabling observations to be made in a tangible manner. This paper describes the use of 3-D printed models of the vocal tract for different vowels that are used to create an acoustic output when stimulated with an appropriate sound source in a new musical instrument: the Vocal Tract Organ. The shape of each printed vocal tract is recovered from magnetic resonance imaging. It sits atop a loudspeaker to which is provided an acoustic L-F model larynx input signal that is controlled by the notes played on a musical instrument digital interface device such as a keyboard. The larynx input is subject to vibrato with extent and frequency adjustable as desired within the ranges usually found for human singing. Polyphonic inputs for choral singing textures can be applied via a single loudspeaker and vocal tract, invoking the approximation of linearity in the voice production system, thereby making multiple vowel stops a possibility while keeping the complexity of the instrument in reasonable check. The Vocal Tract Organ offers a much more human and natural sounding result than the traditional Vox Humana stops found in larger pipe organs, offering the possibility of enhancing pipe organs of the future as well as becoming the basis for a "multi-vowel" chamber organ in its own right. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  1. Context, Contrast, and Tone of Voice in Auditory Sarcasm Perception.

    PubMed

    Voyer, Daniel; Thibodeau, Sophie-Hélène; Delong, Breanna J

    2016-02-01

    Four experiments were conducted to investigate the interplay between context and tone of voice in the perception of sarcasm. These experiments emphasized the role of contrast effects in sarcasm perception exclusively by means of auditory stimuli whereas most past research has relied on written material. In all experiments, a positive or negative computer-generated context spoken in a flat emotional tone was followed by a literally positive statement spoken in a sincere or sarcastic tone of voice. Participants indicated for each statement whether the intonation was sincere or sarcastic. In Experiment 1, a congruent context/tone of voice pairing (negative/sarcastic, positive/sincere) produced fast response times and proportions of sarcastic responses in the direction predicted by the tone of voice. Incongruent pairings produced mid-range proportions and slower response times. Experiment 2 introduced ambiguous contexts to determine whether a lower context/statements contrast would affect the proportion of sarcastic responses and response time. Results showed the expected findings for proportions (values between those obtained for congruent and incongruent pairings in the direction predicted by the tone of voice). However, response time failed to produce the predicted pattern, suggesting potential issues with the choice of stimuli. Experiments 3 and 4 extended the results of Experiments 1 and 2, respectively, to auditory stimuli based on written vignettes used in neuropsychological assessment. Results were exactly as predicted by contrast effects in both experiments. Taken together, the findings suggest that both context and tone influence how sarcasm is perceived while supporting the importance of contrast effects in sarcasm perception.

  2. An exploratory study of voice change associated with healthy speakers after transcutaneous electrical stimulation to laryngeal muscles.

    PubMed

    Fowler, Linda P; Gorham-Rowan, Mary; Hapner, Edie R

    2011-01-01

    The purpose of this study was to determine if measurable changes in fundamental frequency (F(0)) and relative sound level (RSL) occurred in healthy speakers after transcutaneous electrical stimulation (TES) as applied via VitalStim (Chattanooga Group, Chattanooga, TN). A prospective, repeated-measures design. Ten healthy female and 10 healthy male speakers, 20-53 years of age, participated in the study. All participants were nonsmokers and reported negative history for voice disorders. Participants received 1 hour of TES while engaged in eating, drinking, and conversation to simulate a typical dysphagia therapy protocol. Voice recordings were obtained before and immediately after TES. The voice samples consisted of a sustained vowel task and reading of the Rainbow Passage. Measurements of F(0) and RSL were obtained using TF32 (Milenkovic, 2005, University of Wisconsin). The participants also reported any sensations 5 minutes and 24 hours after TES. Measurable changes in F(0) and RSL were found for both tasks but were variable in direction and magnitude. These changes were not statistically significant. Subjective comments ranged from reports of a vocal warm-up feeling to delayed onset muscle soreness. These findings demonstrate that application of TES produces measurable changes in F(0) and RSL. However, the direction and magnitude of these changes are highly variable. Further research is needed to determine factors that may affect the extent to which TES contributes to significant changes in voice. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  3. The influence of visual feedback and register changes on sign language production: A kinematic study with deaf signers

    PubMed Central

    EMMOREY, KAREN; GERTSBERG, NELLY; KORPICS, FRANCO; WRIGHT, CHARLES E.

    2009-01-01

    Speakers monitor their speech output by listening to their own voice. However, signers do not look directly at their hands and cannot see their own face. We investigated the importance of a visual perceptual loop for sign language monitoring by examining whether changes in visual input alter sign production. Deaf signers produced American Sign Language (ASL) signs within a carrier phrase under five conditions: blindfolded, wearing tunnel-vision goggles, normal (citation) signing, shouting, and informal signing. Three-dimensional movement trajectories were obtained using an Optotrak Certus system. Informally produced signs were shorter with less vertical movement. Shouted signs were displaced forward and to the right and were produced within a larger volume of signing space, with greater velocity, greater distance traveled, and a longer duration. Tunnel vision caused signers to produce less movement within the vertical dimension of signing space, but blind and citation signing did not differ significantly on any measure, except duration. Thus, signers do not “sign louder” when they cannot see themselves, but they do alter their sign production when vision is restricted. We hypothesize that visual feedback serves primarily to fine-tune the size of signing space rather than as input to a comprehension-based monitor. PMID:20046943

  4. The influence of visual feedback and register changes on sign language production: A kinematic study with deaf signers.

    PubMed

    Emmorey, Karen; Gertsberg, Nelly; Korpics, Franco; Wright, Charles E

    2009-01-01

    Speakers monitor their speech output by listening to their own voice. However, signers do not look directly at their hands and cannot see their own face. We investigated the importance of a visual perceptual loop for sign language monitoring by examining whether changes in visual input alter sign production. Deaf signers produced American Sign Language (ASL) signs within a carrier phrase under five conditions: blindfolded, wearing tunnel-vision goggles, normal (citation) signing, shouting, and informal signing. Three-dimensional movement trajectories were obtained using an Optotrak Certus system. Informally produced signs were shorter with less vertical movement. Shouted signs were displaced forward and to the right and were produced within a larger volume of signing space, with greater velocity, greater distance traveled, and a longer duration. Tunnel vision caused signers to produce less movement within the vertical dimension of signing space, but blind and citation signing did not differ significantly on any measure, except duration. Thus, signers do not "sign louder" when they cannot see themselves, but they do alter their sign production when vision is restricted. We hypothesize that visual feedback serves primarily to fine-tune the size of signing space rather than as input to a comprehension-based monitor.

  5. "You Need to Let Your Voice Be Heard": Research Participants' Views on Research

    ERIC Educational Resources Information Center

    McDonald, K. E.; Kidney, C. A.; Patka, M.

    2013-01-01

    Background: Persons with intellectual and developmental disabilities have had regrettably few opportunities to voice their opinions on aspects of research with which they have had direct experience. Understanding and responding to these views can contribute to policies and practices that increasingly treat people as they desire to be treated.…

  6. Children's Recognition of Their Own Recorded Voice: Influence of Age and Phonological Impairment

    ERIC Educational Resources Information Center

    Strombergsson, Sofia

    2013-01-01

    Children with phonological impairment (PI) often have difficulties perceiving insufficiencies in their own speech. The use of recordings has been suggested as a way of directing the child's attention toward his/her own speech, despite a lack of evidence that children actually recognize their recorded voice as their own. We present two studies of…

  7. Advanced Electronic Technology

    DTIC Science & Technology

    1977-11-15

    Electronics 15 III. Materials Research 15 TV. Microelectronics 16 V. Surface- Wave Technology 16 DATA SYSTEMS DIVISION 2 INTRODUCTION This...Processing Digital Voice Processing Packet Speech Wideband Integrated Voice/Data Technology Radar Signal Processing Technology Nuclear Safety Designs...facilities make it possible to track the status of these jobs, retrieve their job control language listings, and direct a copy of printed or punched

  8. Voice amplification versus vocal hygiene instruction for teachers with voice disorders: a treatment outcomes study.

    PubMed

    Roy, Nelson; Weinrich, Barbara; Gray, Steven D; Tanner, Kristine; Toledo, Sue Walker; Dove, Heather; Corbin-Lewis, Kim; Stemple, Joseph C

    2002-08-01

    Voice problems are common among schoolteachers. This prospective, randomized clinical trial used patient-based treatment outcomes measures combined with acoustic analysis to evaluate the effectiveness of two treatment programs. Forty-four voice-disordered teachers were randomly assigned to one of three groups: voice amplification using the ChatterVox portable amplifier (VA, n = 15), vocal hygiene (VH, n = 15), and a nontreatment control group (n = 14). Before and after a 6-week treatment phase, all teachers completed: (a) the Voice Handicap Index (VHI), an instrument designed to appraise the self-perceived psychosocial consequences of voice disorders; (b) a voice severity self-rating scale; and (c) an audiorecording for later acoustic analysis. Based on pre- and posttreatment comparisons, only the amplification group experienced significant reductions on mean VHI scores (p = .045), voice severity self-ratings (p = .012), and the acoustic measures of percent jitter (p = .031) and shimmer (p = .008). The nontreatment control group reported a significant increase in level of vocal handicap as assessed by the VHI (p = .012). Although most pre- to posttreatment changes were in the desired direction, no significant improvements were observed within the VH group on any of the dependent measures. Between-group comparisons involving the three possible pairings of the groups revealed a pattern of results to suggest that: (a) compared to the control group, both treatment groups (i.e., VA and VH) experienced significantly more improvement on specific outcomes measures and (b) there were no significant differences between the VA and VH groups to indicate superiority of one treatment over another. Results, however, from a posttreatment questionnaire regarding the perceived benefits of treatment revealed that, compared to the VH group, the VA group reported more clarity of their speaking and singing voice (p = .061), greater ease of voice production (p = .001), and greater compliance with the treatment program (p = .045). These findings clearly support the clinical utility of voice amplification as an alternative for the treatment of voice problems in teachers.

  9. Development of a miniature multiple reference optical coherence tomography imaging device

    NASA Astrophysics Data System (ADS)

    McNamara, Paul M.; O'Riordan, Colm; Collins, Seán.; O'Brien, Peter; Wilson, Carol; Hogan, Josh; Leahy, Martin J.

    2016-03-01

    Multiple reference optical coherence tomography (MR-OCT) is a new technology ideally suited to low-cost, compact OCT imaging. This modality is an extension of time-domain OCT with the addition of a partial mirror in front of the reference mirror. This enables extended, simultaneous depth scanning with the relatively short sweep of a miniature voice coil motor on which the scanning mirror is mounted. Applications of this technology include biometric security, ophthalmology, personal health monitoring and non-destructive testing. This work details early-stage development of the first iteration of a miniature MR-OCT device. This device utilizes a fiber-coupled input from an off-board superluminescent diode (SLD). Typical dimensions of the module are 40 × 57 mm, but future designs are expected to be more compact. Off-the-shelf miniature optical components, voice coil motors and photodetectors are used, with the complexity of design depending on specific applications. The photonic module can be configured as either polarized or non-polarized and can include balanced detection. The photodetectors are directly connected to a printed circuit board under the module containing a transimpedance amplifier with complimentary outputs. The results shown in this work are from the non-polarized device. Assembly of the photonic modules requires extensive planning. In choosing the optical components, Zemax simulations are performed to model the beam characteristics. The physical layout is modeled using Solidworks and each component is placed and aligned via a well-designed alignment procedure involving an active-alignment pick-and-place assembly system.

  10. Replacing Voice Input with Technology that Provided Immediate Visual and Audio Feedback to Reduce Employee Errors

    ERIC Educational Resources Information Center

    Goomas, David T.

    2010-01-01

    In this report from the field at two auto parts distribution centers, order selectors picked auto accessories (e.g., fuses, oil caps, tool kits) into industrial plastic totes as part of store orders. Accurately identifying all store order totes via the license plate number was a prerequisite for the warehouse management system (WMS) to track each…

  11. Portuguese Adaptation and Input for the Validation of the Views on Inpatient Care (VOICE) Outcome Measure to Assess Service Users'Perceptions of Inpatient Psychiatric Care.

    PubMed

    Palha, João; Palha, Filipa; Dias, Pedro; Gonçalves-Pereira, Manuel

    2017-11-29

    Patient satisfaction is an important measure of health care quality. Patients' views have seldom been considered in the construction of measures addressing satisfaction with inpatient facilities in psychiatry. The Views on Inpatient Care - VOICE - is a first service-user generated outcome measure relying solely on their perceptions of acute care, representing a valuable indicator of service users' perceived quality of care. The present study aimed to contribute to the validation of the Portuguese version of VOICE. The questionnaire was translated into Portuguese and applied to a sample of eighty-five female inpatients of a psychiatric institution. Data analysis focused on assessing reliability and exploring the impact of demographic and clinical variables on participants' satisfaction. Internal consistency of the questionnaire was high (α = 0.87). Participants' age and marital status were associated with differences in scores, with older patients and patients who were married or involved in a close relationship presenting higher satisfaction levels. The questionnaire demonstrated good internal consistency and acceptability, as well as construct validity. Further studies should expand the analysis of the psychometric properties of this measure e.g., test-retest reliability. The Portuguese version of VOICE is a promising tool to assess service users' perceptions of inpatient psychiatric care in Portugal.

  12. Voice-enabled Knowledge Engine using Flood Ontology and Natural Language Processing

    NASA Astrophysics Data System (ADS)

    Sermet, M. Y.; Demir, I.; Krajewski, W. F.

    2015-12-01

    The Iowa Flood Information System (IFIS) is a web-based platform developed by the Iowa Flood Center (IFC) to provide access to flood inundation maps, real-time flood conditions, flood forecasts, flood-related data, information and interactive visualizations for communities in Iowa. The IFIS is designed for use by general public, often people with no domain knowledge and limited general science background. To improve effective communication with such audience, we have introduced a voice-enabled knowledge engine on flood related issues in IFIS. Instead of navigating within many features and interfaces of the information system and web-based sources, the system provides dynamic computations based on a collection of built-in data, analysis, and methods. The IFIS Knowledge Engine connects to real-time stream gauges, in-house data sources, analysis and visualization tools to answer natural language questions. Our goal is the systematization of data and modeling results on flood related issues in Iowa, and to provide an interface for definitive answers to factual queries. The goal of the knowledge engine is to make all flood related knowledge in Iowa easily accessible to everyone, and support voice-enabled natural language input. We aim to integrate and curate all flood related data, implement analytical and visualization tools, and make it possible to compute answers from questions. The IFIS explicitly implements analytical methods and models, as algorithms, and curates all flood related data and resources so that all these resources are computable. The IFIS Knowledge Engine computes the answer by deriving it from its computational knowledge base. The knowledge engine processes the statement, access data warehouse, run complex database queries on the server-side and return outputs in various formats. This presentation provides an overview of IFIS Knowledge Engine, its unique information interface and functionality as an educational tool, and discusses the future plans for providing knowledge on flood related issues and resources. IFIS Knowledge Engine provides an alternative access method to these comprehensive set of tools and data resources available in IFIS. Current implementation of the system accepts free-form input and voice recognition capabilities within browser and mobile applications.

  13. Voice/Data Integration in Mobile Radio Networks: Overview and Future Research Directions

    DTIC Science & Technology

    1989-09-30

    degradation in interactive speech when delays are less than about 300 ms (Gold 1977; Gitman and Frank, 1978). When delays are larger (between 300 ms and 1.5...222-267. Gitman , 1. and H. Frank (1978), "Economic Analysis of Integrated Voice and Data Networks: A Case Study," Proc. IEEE 66 1549-1570. Glynn, P.W

  14. "Who's in Charge Here?": Teaching Narrative Voice in Frank O'Connor's "My Oedipus Complex."

    ERIC Educational Resources Information Center

    Wentworth, Michael

    2001-01-01

    Considers how Frank O'Connor's "My Oedipus Complex" provides a good introduction to the subtleties of narrative voice and control. Concludes by considering the notion of control and its relation to the narrative point of view in O'Connor's story and how it bears directly upon the value of reading literature and the reader's role. (SG)

  15. The Effects of Rate of Deviation and Musical Context on Intonation Perception in Homophonic Four-Part Chorales.

    NASA Astrophysics Data System (ADS)

    Bell, Michael Stephen

    Sixty-four trained musicians listened to four -bar excerpts of selected chorales by J. S. Bach, which were presented both in four-part texture (harmonic context) and as a single voice part (melodic context). These digitally synthesized examples were created by combining the first twelve partials, and all voice parts had the same generic timbre. A within-subjects design was used, so subjects heard each example in both contexts. Included in the thirty -two excerpts for each subject were four soprano, four alto, four tenor, and four bass parts as the target voices. The intonation of the target voice was varied such that the voice stayed in tune or changed by a half cent, two cents, or eight cents per second (a cent is 1/100 of a half step). Although direction of the deviation (sharp or flat) was not a significant factor in intonation perception, main effects for context (melodic vs. harmonic) and rate of deviation were highly significant, as was the interaction between rate of deviation and context. Specifically, selections that stayed in tune or changed only by half cents were not perceived differently; for larger deviations, the error was detected earlier and the intonation was judged to be worse in the harmonic contexts compared to the melodic contexts. Additionally, the direction of the error was correctly identified in the melodic context more often than the hamonic context only for the examples that mistuned at a rate of eight cents per second. Correct identification of the voice part that went out of tune in the four-part textures depended only on rate of deviation: the in tune excerpts (no voice going out of tune) and the eight cent deviations were correctly identified most often, the two cent deviations were next, and the half cent deviation excerpts were the least accurately identified.

  16. A framework for the design of a voice-activated, intelligent, and hypermedia-based aircraft maintenance manual

    NASA Astrophysics Data System (ADS)

    Patankar, Manoj Shashikant

    Federal Aviation Regulations require Aviation Maintenance Technicians (AMTs) to refer to approved maintenance manuals when performing maintenance on airworthy aircraft. Because these manuals are paper-based, larger the size of the aircraft, more cumbersome are the manuals. Federal Aviation Administration (FAA) recognized the difficulties associated with the use of large manuals and conducted studies on the use of electronic media as an alternative to the traditional paper format. However, these techniques do not employ any artificial intelligence technologies and the user interface is limited to either a keyboard or a stylus pen. The primary emphasis of this research was to design a generic framework that would allow future development of voice-activated, intelligent, and hypermedia-based aircraft maintenance manuals. A prototype (VIHAMS-Voice-activated, Intelligent, and Hypermedia-based Aircraft Maintenance System) was developed, as a secondary emphasis, using the design and development techniques that evolved from this research. An evolutionary software design approach was used to design the proposed framework and the structured rapid prototyping technique was used to produce the VIHAMS prototype. VoiceAssist by Creative Labs was used to provide the voice interface so that the users (AMTs) could keep their hands free to work on the aircraft while maintaining complete control over the computer through discrete voice commands. KnowledgePro for Windows sp{TM}, an expert system shell, provided "intelligence" to the prototype. As a result of this intelligence, the system provided expert guidance to the user. The core information contained in conventional manuals was available in a hypermedia format. The prototype's operating hardware included a notebook computer with a fully functional audio system. An external microphone and the built-in speaker served as the input and output devices (along with the color monitor), respectively. Federal Aviation Administration estimates the United States air carriers to operate 3,991 large jet aircraft in the year 1996 (FAA Aviation Forecasts, 1987-1998). With an estimate of seventy manuals per such aircraft, the development of intelligent manuals is expected to impact 279,370 manuals in this country. Soon, over 55 thousand maintenance technicians will be able to carry the seven pound system to an aircraft, use voice commands to access the aircraft's files on the system, seek assistance from the expert system to diagnose the fault, and obtain instructions on how to rectify the fault. The evolutionary design approach and the rapid prototyping techniques were very well suited for the spiral testing strategy. Therefore, this strategy was used to test the structural and functional validity of this research. Professors Darrell Anderson and Brian Stout (Aviation faculty at San Jose State University) and Mr. Gregory Shea (a United Airlines mechanic and SJSU student) are representatives of the real-world users of the final product. Therefore, they conducted the alpha test of this prototype. Mr. Daniel Neal and Mr. Stephen Harms have been actively involved in light aircraft maintenance for more than ten years. They evaluated the prototype's usability. All the above evaluators used standard testing tools and evaluated the prototype under field conditions. The evaluators concluded that the VIHAMS prototype used a valid fault diagnosis strategy, the system architecture could be used to develop similar systems using off-the-shelf tools, and the voice input system could be refined to improve its usability.

  17. Relationship between self-focused attention, mindfulness and distress in individuals with auditory verbal hallucinations.

    PubMed

    Úbeda-Gómez, J; León-Palacios, M G; Escudero-Pérez, S; Barros-Albarrán, M D; López-Jiménez, A M; Perona-Garcelán, S

    2015-01-01

    The purpose of this study was to investigate the relationships among self-focused attention, mindfulness and distress caused by the voices in psychiatric patients. Fifty-one individuals with a psychiatric diagnosis participated in this study. The Psychotic Symptom Rating Scale (PSYRATS) emotional factor was applied to measure the distress caused by the voices, the Self-Absorption Scale (SAS) was given for measuring the levels of self-focused attention, and the Mindful Attention Awareness Scale (MAAS) was used to measure mindfulness. The results showed that distress caused by the voices correlated positively with self-focused attention (private and public) and negatively with mindfulness. A negative correlation was also found between mindfulness and self-focused attention (private and public). Finally, multiple linear regression analysis showed that public self-focus was the only factor predicting distress caused by the voices. Intervention directed at diminishing public self-focused attention and increasing mindfulness could improve distress caused by the voices.

  18. Using Rate of Divergence as an Objective Measure to Differentiate between Voice Signal Types Based on the Amount of Disorder in the Signal.

    PubMed

    Calawerts, William M; Lin, Liyu; Sprott, J C; Jiang, Jack J

    2017-01-01

    The purpose of this paper is to introduce the rate of divergence as an objective measure to differentiate between the four voice types based on the amount of disorder present in a signal. We hypothesized that rate of divergence would provide an objective measure that can quantify all four voice types. A total of 150 acoustic voice recordings were randomly selected and analyzed using traditional perturbation, nonlinear, and rate of divergence analysis methods. We developed a new parameter, rate of divergence, which uses a modified version of Wolf's algorithm for calculating Lyapunov exponents of a system. The outcome of this calculation is not a Lyapunov exponent, but rather a description of the divergence of two nearby data points for the next three points in the time series, followed in three time-delayed embedding dimensions. This measure was compared to currently existing perturbation and nonlinear dynamic methods of distinguishing between voice signals. There was a direct relationship between voice type and rate of divergence. This calculation is especially effective at differentiating between type 3 and type 4 voices (P < 0.001) and is equally effective at differentiating type 1, type 2, and type 3 signals as currently existing methods. The rate of divergence calculation introduced is an objective measure that can be used to distinguish between all four voice types based on the amount of disorder present, leading to quicker and more accurate voice typing as well as an improved understanding of the nonlinear dynamics involved in phonation. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  19. Using rate of divergence as an objective measure to differentiate between voice signal types based on the amount of disorder in the signal

    PubMed Central

    Calawerts, William M; Lin, Liyu; Sprott, JC; Jiang, Jack J

    2016-01-01

    Objective/Hypothesis The purpose of this paper is to introduce rate of divergence as an objective measure to differentiate between the four voice types based on the amount of disorder present in a signal. We hypothesized that rate of divergence would provide an objective measure that can quantify all four voice types. Study Design 150 acoustic voice recordings were randomly selected and analyzed using traditional perturbation, nonlinear, and rate of divergence analysis methods. ty Methods We developed a new parameter, rate of divergence, which uses a modified version of Wolf’s algorithm for calculating Lyapunov exponents of a system. The outcome of this calculation is not a Lyapunov exponent, but rather a description of the divergence of two nearby data points for the next three points in the time series, followed in three time delayed embedding dimensions. This measure was compared to currently existing perturbation and nonlinear dynamic methods of distinguishing between voice signals. Results There was a direct relationship between voice type and rate of divergence. This calculation is especially effective at differentiating between type 3 and type 4 voices (p<0.001), and is equally effective at differentiating type 1, type 2, and type 3 signals as currently existing methods. Conclusion The rate of divergence calculation introduced is an objective measure that can be used to distinguish between all four voice types based on amount of disorder present, leading to quicker and more accurate voice typing as well as an improved understanding of the nonlinear dynamics involved in phonation. PMID:26920858

  20. Vocal parameters and voice-related quality of life in adult women with and without ovarian function.

    PubMed

    Ferraz, Pablo Rodrigo Rocha; Bertoldo, Simão Veras; Costa, Luanne Gabrielle Morais; Serra, Emmeliny Cristini Nogueira; Silva, Eduardo Magalhães; Brito, Luciane Maria Oliveira; Chein, Maria Bethânia da Costa

    2013-05-01

    To identify the perceptual and acoustic parameters of voice in adult women with and without ovarian function and its impact on quality of life related to voice. Cross-sectional and analytical study with 106 women divided into, two groups: G1, with ovarian function (n=43) and G2, without physiological ovarian function (n=63). The women were instructed to sustain the vowel "a" and the sounds of /s/ and /z/ in habitual pitch and loudness. They were also asked to classify their voices and answer the voice-related quality of life (V-RQOL) questionnaire. The perceptual analysis of the vocal samples was performed by three speech-language pathologists using the GRBASI (G: grade; R: roughness; B: breathness; A: asthenia; S: strain; I: instability) scale. The acoustic analysis was carried out with the software VoxMetria 2.7h (CTS Informatica). The data were analyzed using descriptive statistics. In the perceptual analysis, both groups showed a mild deviation for the parameters roughness, strain, and instability, but only G2 showed a mild impact for the overall degree of dysphonia. The mean of fundamental frequency was significantly lower for the G2, with a difference of 17.41Hz between the two groups. There was no impact on V-RQOL in any of the V-RQOL domains for this group. With the menopause, there is a change in women's voices, impacting on some voice parameters. However, there is no direct impact on their quality of life related to voice. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  1. Data compression/error correction digital test system. Appendix 2: Theory of operation

    NASA Technical Reports Server (NTRS)

    1972-01-01

    An overall block diagram of the DC/EC digital system test is shown. The system is divided into two major units: the transmitter and the receiver. In operation, the transmitter and receiver are connected only by a real or simulated transmission link. The system inputs consist of: (1) standard format TV video, (2) two channels of analog voice, and (3) one serial PCM bit stream.

  2. Criteria for Appraising Computer-Based Simulations for Teaching Arabic as a Foreign Language

    DTIC Science & Technology

    2005-04-01

    activity abroad that most contributed to their increase in fluency was ‘hanging out’ with Russian friends, defined as visiting, eating, and watching...approach is testing that learning has indeed occurred, in that a teacher must evaluate not only linguistic accuracy but also fluency in the proper...written responses, with student input analyzed using voice processing technology. Cultural Proficiency in Arabic Fluency in a foreign language

  3. Computer simulator for a mobile telephone system

    NASA Technical Reports Server (NTRS)

    Schilling, D. L.

    1981-01-01

    A software simulator was developed to assist NASA in the design of the land mobile satellite service. Structured programming techniques were used by developing the algorithm using an ALCOL-like pseudo language and then encoding the algorithm into FORTRAN 4. The basic input data to the system is a sine wave signal although future plans call for actual sampled voice as the input signal. The simulator is capable of studying all the possible combinations of types and modes of calls through the use of five communication scenarios: single hop systems; double hop, signal gateway system; double hop, double gateway system; mobile to wireline system; and wireline to mobile system. The transmitter, fading channel, and interference source simulation are also discussed.

  4. A human factors approach to range scheduling for satellite control

    NASA Technical Reports Server (NTRS)

    Wright, Cameron H. G.; Aitken, Donald J.

    1991-01-01

    Range scheduling for satellite control presents a classical problem: supervisory control of a large-scale dynamic system, with unwieldy amounts of interrelated data used as inputs to the decision process. Increased automation of the task, with the appropriate human-computer interface, is highly desirable. The development and user evaluation of a semi-automated network range scheduling system is described. The system incorporates a synergistic human-computer interface consisting of a large screen color display, voice input/output, a 'sonic pen' pointing device, a touchscreen color CRT, and a standard keyboard. From a human factors standpoint, this development represents the first major improvement in almost 30 years to the satellite control network scheduling task.

  5. Covert digital manipulation of vocal emotion alter speakers’ emotional states in a congruent direction

    PubMed Central

    Johansson, Petter; Hall, Lars; Segnini, Rodrigo; Mercadié, Lolita; Watanabe, Katsumi

    2016-01-01

    Research has shown that people often exert control over their emotions. By modulating expressions, reappraising feelings, and redirecting attention, they can regulate their emotional experience. These findings have contributed to a blurring of the traditional boundaries between cognitive and emotional processes, and it has been suggested that emotional signals are produced in a goal-directed way and monitored for errors like other intentional actions. However, this interesting possibility has never been experimentally tested. To this end, we created a digital audio platform to covertly modify the emotional tone of participants’ voices while they talked in the direction of happiness, sadness, or fear. The result showed that the audio transformations were being perceived as natural examples of the intended emotions, but the great majority of the participants, nevertheless, remained unaware that their own voices were being manipulated. This finding indicates that people are not continuously monitoring their own voice to make sure that it meets a predetermined emotional target. Instead, as a consequence of listening to their altered voices, the emotional state of the participants changed in congruence with the emotion portrayed, which was measured by both self-report and skin conductance level. This change is the first evidence, to our knowledge, of peripheral feedback effects on emotional experience in the auditory domain. As such, our result reinforces the wider framework of self-perception theory: that we often use the same inferential strategies to understand ourselves as those that we use to understand others. PMID:26755584

  6. Using quality function deployment to capture the voice of the customer and translate it into the voice of the provider.

    PubMed

    Chaplin, E; Bailey, M; Crosby, R; Gorman, D; Holland, X; Hippe, C; Hoff, T; Nawrocki, D; Pichette, S; Thota, N

    1999-06-01

    Health care has a number of historical barriers to capturing the voice of the customer and to incorporating customer wants into health care services, whether the customer is a patient, an insurer, or a community. Quality function deployment (QFD) is a set of tools and practices that can help overcome these barriers to form a process for the planning and design or redesign of products and services. The goal of the project was to increase referral volume and to improve a rehabilitation hospital's capacity to provide comprehensive medical and/or legal evaluations for people with complex and catastrophic injuries or illnesses. HIGH-LEVEL VIEW OF QFD AS A PROCESS: The steps in QFD are as follows: capture of the voice of the customer, quality deployment, functions deployment, failure mode deployment, new process deployment, and task deployment. The output of each step becomes the input to a matrix tool or table of the next step of the process. In 3 1/2 months a nine-person project team at Continental Rehabilitation Hospital (San Diego) used QFD tools to capture the voice of the customer, use these data as the basis for a questionnaire on important qualities of service from the customer's perspective, obtain competitive data on how the organization was perceived to be meeting the demanded qualities, identify measurable dimensions and targets of these qualities, and incorporate the functions and tasks into the delivery of service which are necessary to meet the demanded qualities. The future of providing health care services will belong to organizations that can adapt to a rapidly changing environment and to demands for new products and services that are produced and delivered in new ways.

  7. Effect of Botulinum Toxin and Surgery among Spasmodic Dysphonia Patients.

    PubMed

    van Esch, Babette F; Wegner, Inge; Stegeman, Inge; Grolman, Wilko

    2017-02-01

    Objective The effect of botulinum toxin among patients with adductor spasmodic dysphonia (AdSD) is temporary. To optimize long-term treatment outcome, other therapy options should be evaluated. Alternative treatment options for AdSD comprise several surgical treatments, such as thyroarytenoid myotomy, thyroplasty, selective laryngeal adductor denervation-reinnervation, laryngeal nerve crush, and recurrent laryngeal nerve resection. Here, we present the first systematic review comparing the effect of botulinum toxin with surgical treatment among patients diagnosed with AdSD. Data Sources MEDLINE (PubMed), EMBASE, and the Cochrane Library. Methods Articles were reviewed by 2 independent authors, and data were compiled in tables for analysis of the objective outcome (voice expert evaluation after voice recording), the subjective outcome (patient self-assessment scores), and voice-related quality of life (Voice Health Index scores). Results No clinical trials comparing both treatment modalities were identified. Single-armed studies evaluated either the effect of botulinum toxin or surgical treatment. Thirteen studies reported outcomes after botulinum toxin treatment (n = 419), and 9 studies reported outcomes after surgical treatment (n = 585 patients). A positive effect of bilateral botulinum toxin injections was found for the objective voice outcome, subjective voice outcome, and quality of life. The duration of the beneficial effect ranged from 15 to 18 weeks. Surgical treatment had an overall positive effect on objective voice improvement, subjective voice improvement, and quality of live. Conclusion No preference for one treatment could be demonstrated. Prospective clinical trials comparing treatment modalities are recommended to delineate the optimal outcomes by direct comparison.

  8. Classification of vocal aging using parameters extracted from the glottal signal.

    PubMed

    Forero Mendoza, Leonardo A; Cataldo, Edson; Vellasco, Marley M B R; Silva, Marco A; Apolinário, José A

    2014-09-01

    This article proposes and evaluates a method to classify vocal aging using artificial neural network (ANN) and support vector machine (SVM), using the parameters extracted from the speech signal as inputs. For each recorded speech, from a corpus of male and female speakers of different ages, the corresponding glottal signal is obtained using an inverse filtering algorithm. The Mel Frequency Cepstrum Coefficients (MFCC) also extracted from the voice signal and the features extracted from the glottal signal are supplied to an ANN and an SVM with a previous selection. The selection is performed by a wrapper approach of the most relevant parameters. Three groups are considered for the aging-voice classification: young (aged 15-30 years), adult (aged 31-60 years), and senior (aged 61-90 years). The results are compared using different possibilities: with only the parameters extracted from the glottal signal, with only the MFCC, and with a combination of both. The results demonstrate that the best classification rate is obtained using the glottal signal features, which is a novel result and the main contribution of this article. Copyright © 2014 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  9. Evaluation of voice codecs for the Australian mobile satellite system

    NASA Technical Reports Server (NTRS)

    Bundrock, Tony; Wilkinson, Mal

    1990-01-01

    The evaluation procedure to choose a low bit rate voice coding algorithm is described for the Australian land mobile satellite system. The procedure is designed to assess both the inherent quality of the codec under 'normal' conditions and its robustness under 'severe' conditions. For the assessment, normal conditions were chosen to be random bit error rate with added background acoustic noise and the severe condition is designed to represent burst error conditions when mobile satellite channel suffers from signal fading due to roadside vegetation. The assessment is divided into two phases. First, a reduced set of conditions is used to determine a short list of candidate codecs for more extensive testing in the second phase. The first phase conditions include quality and robustness and codecs are ranked with a 60:40 weighting on the two. Second, the short listed codecs are assessed over a range of input voice levels, BERs, background noise conditions, and burst error distributions. Assessment is by subjective rating on a five level opinion scale and all results are then used to derive a weighted Mean Opinion Score using appropriate weights for each of the test conditions.

  10. Tracheo-bronchial soft tissue and cartilage resonances in the subglottal acoustic input impedance.

    PubMed

    Lulich, Steven M; Arsikere, Harish

    2015-06-01

    This paper offers a re-evaluation of the mechanical properties of the tracheo-bronchial soft tissues and cartilage and uses a model to examine their effects on the subglottal acoustic input impedance. It is shown that the values for soft tissue elastance and cartilage viscosity typically used in models of subglottal acoustics during phonation are not accurate, and corrected values are proposed. The calculated subglottal acoustic input impedance using these corrected values reveals clusters of weak resonances due to soft tissues (SgT) and cartilage (SgC) lining the walls of the trachea and large bronchi, which can be observed empirically in subglottal acoustic spectra. The model predicts that individuals may exhibit SgT and SgC resonances to variable degrees, depending on a number of factors including tissue mechanical properties and the dimensions of the trachea and large bronchi. Potential implications for voice production and large pulmonary airway tissue diseases are also discussed.

  11. Neuroprosthetics and the science of patient input

    PubMed Central

    Civillico, Eugene F.

    2017-01-01

    Safe and effective neuroprosthetic systems are of great interest to both DARPA and CDRH, due to their innovative nature and their potential to aid severely disabled populations. By expanding what is possible in human-device interaction, these devices introduce new potential benefits and risks. Therefore patient input, which is increasingly important in weighing benefits and risks, is particularly relevant for this class of devices. FDA has been a significant contributor to an ongoing stakeholder conversation about the inclusion of the patient voice, working collaboratively to create a new framework for a patient-centered approach to medical device development. This framework is evolving through open dialogue with researcher and patient communities, investment in the science of patient input, and policymaking that is responsive to patient-centered data throughout the total product life cycle. In this commentary, we will discuss recent developments in patient-centered benefit-risk assessment and their relevance to the development of neural prosthetic systems. PMID:27456271

  12. Neuroprosthetics and the science of patient input.

    PubMed

    Benz, Heather L; Civillico, Eugene F

    2017-01-01

    Safe and effective neuroprosthetic systems are of great interest to both DARPA and CDRH, due to their innovative nature and their potential to aid severely disabled populations. By expanding what is possible in human-device interaction, these devices introduce new potential benefits and risks. Therefore patient input, which is increasingly important in weighing benefits and risks, is particularly relevant for this class of devices. FDA has been a significant contributor to an ongoing stakeholder conversation about the inclusion of the patient voice, working collaboratively to create a new framework for a patient-centered approach to medical device development. This framework is evolving through open dialogue with researcher and patient communities, investment in the science of patient input, and policymaking that is responsive to patient-centered data throughout the total product life cycle. In this commentary, we will discuss recent developments in patient-centered benefit-risk assessment and their relevance to the development of neural prosthetic systems. Published by Elsevier Inc.

  13. High or low? Comparing high and low-variability phonetic training in adult and child second language learners

    PubMed Central

    Brown, Helen; Clayards, Meghan

    2017-01-01

    Background High talker variability (i.e., multiple voices in the input) has been found effective in training nonnative phonetic contrasts in adults. A small number of studies suggest that children also benefit from high-variability phonetic training with some evidence that they show greater learning (more plasticity) than adults given matched input, although results are mixed. However, no study has directly compared the effectiveness of high versus low talker variability in children. Methods Native Greek-speaking eight-year-olds (N = 52), and adults (N = 41) were exposed to the English /i/-/ɪ/ contrast in 10 training sessions through a computerized word-learning game. Pre- and post-training tests examined discrimination of the contrast as well as lexical learning. Participants were randomly assigned to high (four talkers) or low (one talker) variability training conditions. Results Both age groups improved during training, and both improved more while trained with a single talker. Results of a three-interval oddity discrimination test did not show the predicted benefit of high-variability training in either age group. Instead, children showed an effect in the reverse direction—i.e., reliably greater improvements in discrimination following single talker training, even for untrained generalization items, although the result is qualified by (accidental) differences between participant groups at pre-test. Adults showed a numeric advantage for high-variability but were inconsistent with respect to voice and word novelty. In addition, no effect of variability was found for lexical learning. There was no evidence of greater plasticity for phonetic learning in child learners. Discussion This paper adds to the handful of studies demonstrating that, like adults, child learners can improve their discrimination of a phonetic contrast via computerized training. There was no evidence of a benefit of training with multiple talkers, either for discrimination or word learning. The results also do not support the findings of greater plasticity in child learners found in a previous paper (Giannakopoulou, Uther & Ylinen, 2013a). We discuss these results in terms of various differences between training and test tasks used in the current work compared with previous literature. PMID:28584698

  14. DIRECT operational field test evaluation natural use study. Part 4, Recommendations for expanded deployment

    DOT National Transportation Integrated Search

    1998-08-01

    The DIRECT project compared four low-cost driver information systems. Of the four that were : compared, the RDS approach proved superior to the others in toggling reliability and voice quality. The DIRECT project planned to expand the implementation ...

  15. Determinants of structural choice in visually situated sentence production.

    PubMed

    Myachykov, Andriy; Garrod, Simon; Scheepers, Christoph

    2012-11-01

    Three experiments investigated how perceptual, structural, and lexical cues affect structural choices during English transitive sentence production. Participants described transitive events under combinations of visual cueing of attention (toward either agent or patient) and structural priming with and without semantic match between the notional verb in the prime and the target event. Speakers had a stronger preference for passive-voice sentences (1) when their attention was directed to the patient, (2) upon reading a passive-voice prime, and (3) when the verb in the prime matched the target event. The verb-match effect was the by-product of an interaction between visual cueing and verb match: the increase in the proportion of passive-voice responses with matching verbs was limited to the agent-cued condition. Persistence of visual cueing effects in the presence of both structural and lexical cues suggests a strong coupling between referent-directed visual attention and Subject assignment in a spoken sentence. Copyright © 2012 Elsevier B.V. All rights reserved.

  16. Comparison of Pitch Strength With Perceptual and Other Acoustic Metric Outcome Measures Following Medialization Laryngoplasty.

    PubMed

    Rubin, Adam D; Jackson-Menaldi, Cristina; Kopf, Lisa M; Marks, Katherine; Skeffington, Jean; Skowronski, Mark D; Shrivastav, Rahul; Hunter, Eric J

    2018-05-14

    The diagnoses of voice disorders, as well as treatment outcomes, are often tracked using visual (eg, stroboscopic images), auditory (eg, perceptual ratings), objective (eg, from acoustic or aerodynamic signals), and patient report (eg, Voice Handicap Index and Voice-Related Quality of Life) measures. However, many of these measures are known to have low to moderate sensitivity and specificity for detecting changes in vocal characteristics, including vocal quality. The objective of this study was to compare changes in estimated pitch strength (PS) with other conventionally used acoustic measures based on the cepstral peak prominence (smoothed cepstral peak prominence, cepstral spectral index of dysphonia, and acoustic voice quality index), and clinical judgments of voice quality (GRBAS [grade, roughness, breathiness, asthenia, strain] scale) following laryngeal framework surgery. This study involved post hoc analysis of recordings from 22 patients pretreatment and post treatment (thyroplasty and behavioral therapy). Sustained vowels and connected speech were analyzed using objective measures (PS, smoothed cepstral peak prominence, cepstral spectral index of dysphonia, and acoustic voice quality index), and these results were compared with mean auditory-perceptual ratings by expert clinicians using the GRBAS scale. All four acoustic measures changed significantly in the direction that usually indicates improved voice quality following treatment (P < 0.005). Grade and breathiness correlated the strongest with the acoustic measures (|r| ~0.7) with strain being the least correlated. Acoustic analysis on running speech highly correlates with judged ratings. PS is a robust, easily obtained acoustic measure of voice quality that could be useful in the clinical environment to follow treatment of voice disorders. Copyright © 2018. Published by Elsevier Inc.

  17. Effects of message, source, and context on evaluations of employee voice behavior.

    PubMed

    Whiting, Steven W; Maynes, Timothy D; Podsakoff, Nathan P; Podsakoff, Philip M

    2012-01-01

    The article contained a production-related error. In Table 5, the four values in the rows for Study 1 Prosocial motives and Study 1 Constructive voice should have been shifted one column to the right, to the Direct and Total Performance evaluations columns. All versions of this article have been corrected.] Although employee voice behavior is expected to have important organizational benefits, research indicates that employees voicing their recommendations for organizational change may be evaluated either positively or negatively by observers. A review of the literature suggests that the perceived efficacy of voice behaviors may be a function of characteristics associated with the (a) source, (b) message, and (c) context of the voice event. In this study, we manipulated variables from each of these categories based on a model designed to predict when voice will positively or negatively impact raters' evaluations of an employee's performance. To test our model, we conducted 3 laboratory studies in which we manipulated 2 source factors (voicer expertise and trustworthiness), 2 message factors (recommending a solution and positively vs. negatively framing the message), and 2 context factors (timing of the voice event and organizational norms for speaking up vs. keeping quiet). We also examined the mediating effects of liking, prosocial motives, and perceptions that the voice behavior was constructive on the relationships between the source, message, and context factors and performance evaluations. Generally speaking, we found that at least one of the variables from each category had an effect on performance evaluations for the voicer and that most of these effects were indirect, operating through one or more of the mediators. Implications for theory and future research are discussed.

  18. Boosting Contextual Information for Deep Neural Network Based Voice Activity Detection

    DTIC Science & Technology

    2015-02-01

    multi-resolution stacking (MRS), which is a stack of ensemble classifiers. Each classifier in a building block inputs the concatenation of the predictions ...a base classifier in MRS, named boosted deep neural network (bDNN). bDNN first generates multiple base predictions from different contexts of a single...frame by only one DNN and then aggregates the base predictions for a better prediction of the frame, and it is different from computationally

  19. DIRECT operational field test evaluation : natural use study : part 2 : driver satisfaction in DIRECT controlling for reliability

    DOT National Transportation Integrated Search

    1998-08-01

    This report describes the DIRECT field test which was designed to evaluate the user benefits, institutional issues, and technical issues of en-route driver advisory and traveler information services. Focus was on testing and evaluating the voice-base...

  20. A miniaturized digital telemetry system for physiological data transmission

    NASA Technical Reports Server (NTRS)

    Portnoy, W. M.; Stotts, L. J.

    1978-01-01

    A physiological date telemetry system, consisting basically of a portable unit and a ground base station was designed, built, and tested. The portable unit to be worn by the subject is composed of a single crystal controlled transmitter with AM transmission of digital data and narrowband FM transmission of voice; a crystal controlled FM receiver; thirteen input channels follwed by a PCM encoder (three of these channels are designed for ECG data); a calibration unit; and a transponder control system. The ground base station consists of a standard telemetry reciever, a decoder, and an FM transmitter for transmission of voice and transponder signals to the portable unit. The ground base station has complete control of power to all subsystems in the portable unit. The phase-locked loop circuit which is used to decode the data, remains in operation even when the signal from the portable unit is interrupted.

  1. V2S: Voice to Sign Language Translation System for Malaysian Deaf People

    NASA Astrophysics Data System (ADS)

    Mean Foong, Oi; Low, Tang Jung; La, Wai Wan

    The process of learning and understand the sign language may be cumbersome to some, and therefore, this paper proposes a solution to this problem by providing a voice (English Language) to sign language translation system using Speech and Image processing technique. Speech processing which includes Speech Recognition is the study of recognizing the words being spoken, regardless of whom the speaker is. This project uses template-based recognition as the main approach in which the V2S system first needs to be trained with speech pattern based on some generic spectral parameter set. These spectral parameter set will then be stored as template in a database. The system will perform the recognition process through matching the parameter set of the input speech with the stored templates to finally display the sign language in video format. Empirical results show that the system has 80.3% recognition rate.

  2. Development of a Voice Activity Controlled Noise Canceller

    PubMed Central

    Abid Noor, Ali O.; Samad, Salina Abdul; Hussain, Aini

    2012-01-01

    In this paper, a variable threshold voice activity detector (VAD) is developed to control the operation of a two-sensor adaptive noise canceller (ANC). The VAD prohibits the reference input of the ANC from containing some strength of actual speech signal during adaptation periods. The novelty of this approach resides in using the residual output from the noise canceller to control the decisions made by the VAD. Thresholds of full-band energy and zero-crossing features are adjusted according to the residual output of the adaptive filter. Performance evaluation of the proposed approach is quoted in terms of signal to noise ratio improvements as well mean square error (MSE) convergence of the ANC. The new approach showed an improved noise cancellation performance when tested under several types of environmental noise. Furthermore, the computational power of the adaptive process is reduced since the output of the adaptive filter is efficiently calculated only during non-speech periods. PMID:22778667

  3. Amygdala and auditory cortex exhibit distinct sensitivity to relevant acoustic features of auditory emotions.

    PubMed

    Pannese, Alessia; Grandjean, Didier; Frühholz, Sascha

    2016-12-01

    Discriminating between auditory signals of different affective value is critical to successful social interaction. It is commonly held that acoustic decoding of such signals occurs in the auditory system, whereas affective decoding occurs in the amygdala. However, given that the amygdala receives direct subcortical projections that bypass the auditory cortex, it is possible that some acoustic decoding occurs in the amygdala as well, when the acoustic features are relevant for affective discrimination. We tested this hypothesis by combining functional neuroimaging with the neurophysiological phenomena of repetition suppression (RS) and repetition enhancement (RE) in human listeners. Our results show that both amygdala and auditory cortex responded differentially to physical voice features, suggesting that the amygdala and auditory cortex decode the affective quality of the voice not only by processing the emotional content from previously processed acoustic features, but also by processing the acoustic features themselves, when these are relevant to the identification of the voice's affective value. Specifically, we found that the auditory cortex is sensitive to spectral high-frequency voice cues when discriminating vocal anger from vocal fear and joy, whereas the amygdala is sensitive to vocal pitch when discriminating between negative vocal emotions (i.e., anger and fear). Vocal pitch is an instantaneously recognized voice feature, which is potentially transferred to the amygdala by direct subcortical projections. These results together provide evidence that, besides the auditory cortex, the amygdala too processes acoustic information, when this is relevant to the discrimination of auditory emotions. Copyright © 2016 Elsevier Ltd. All rights reserved.

  4. Collaboration and conquest: MTD as viewed by voice teacher (singing voice specialist) and speech-language pathologist.

    PubMed

    Goffi-Fynn, Jeanne C; Carroll, Linda M

    2013-05-01

    This study was designed as a qualitative case study to demonstrate the process of diagnosis and treatment between a voice team to manage a singer diagnosed with muscular tension dysphonia (MTD). Traditionally, literature suggests that MTD is challenging to treat and little in the literature directly addresses singers with MTD. Data collected included initial medical screening with laryngologist, referral to speech-language pathologist (SLP) specializing in voice disorders among singers, and adjunctive voice training with voice teacher trained in vocology (singing voice specialist or SVS). Initial target goals with SLP included reducing extrinsic laryngeal tension, using a relaxed laryngeal posture, and effective abdominal-diaphragmatic support for all phonation events. Balance of respiratory forces, laryngeal coordination, and use of optimum filtering of the source signal through resonance and articulatory awareness was emphasized. Further work with SVS included three main goals including a lowered breathing pattern to aid in decreasing subglottic air pressure, vertical laryngeal position to lower to allow for a relaxed laryngeal position, and a top-down singing approach to encourage an easier, more balanced registration, and better resonance. Initial results also emphasize the retraining of subject toward a sensory rather than auditory mode of monitoring. Other areas of consideration include singers' training and vocal use, the psychological effects of MTD, the personalities potentially associated with it, and its relationship with stress. Finally, the results emphasize that a positive rapport with the subject and collaboration between all professionals involved in a singer's care are essential for recovery. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  5. Reliability and Validity of the Turkish Version of the Voice-Related Quality of Life Measure.

    PubMed

    Tezcaner, Zahide Çiler; Aksoy, Songül

    2017-03-01

    This study aims to test the validity and reliability of the Turkish version of the Voice-Related Quality of Life (V-RQOL) questionnaire. This is a nonrandomized, prospective study with control group. The questionnaire was administered to 249 individuals-130 with vocal complaint and 119 without-with a mean age of 37.8 ± 12.3 years. The Turkish version of the Voice Handicap Index (VHI) and perceptual voice evaluation measures were also administered at 2-14 days for retest reliability. The instrument was submitted to validity and reliability evaluation. The V-RQOL measure showed a strong internal consistency and test-retest reliability; the Cronbach's alpha coefficient for the overall V-RQOL was 0.969, the physical functioning domain was 0.949, and the social-emotional domain was 0.940. In the test-retest reliability test, the overall V-RQOL was found to be 0.989. The construct validity of the V-RQOL was determined based on the strength and direction of its relation to the VHI and the perceptual voice evaluation measure. The higher the VHI level, the lower the physical functioning, social-emotional, and overall score levels of the V-RQOL (r = -0.927, r = -0.912, r = -0.944, respectively; P < 0.001). Following the perceptual voice self-assessment, a statistically significant difference was found between the V-RQOL scores of individuals who defined their voices as good, very good, and perfect, and those who defined their voices as bad and very bad (P < 0.001). The results suggest that the Turkish version of the V-RQOL measure has reliability and validity and may play a crucial role in evaluating Turkish-speaking patients with voice disorders. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  6. Speak Up Speak Out Coalition Survey Results | Science ...

    EPA Pesticide Factsheets

    Comprehensive planning is a visionary planning process that integrates community values and land use policy. The Mayor of Duluth, Minnesota, directed the inclusion of two new values into the City’s comprehensive planning process to direct the community’s future, process: health and fairness. In order to understand the meanings of health and fairness that residents of the city hold, the Community Planning Department included questions in a city-wide survey of planning priorities. As a community organization reviewed the survey results that would inform the new directives, they realized that overburdened communities were underrepresented in the survey responses. To address this deficiency, the community organization asked the City of Duluth if they could conduct a survey of the underrepresented voices to ensure their input was included in the process. The Health in All Policies Coalition contacted the USEPA Office of Research and Development in Duluth, MN at the advice of the Planning Department. The support USEPA provided ensured that the Coalition could make recommendations to the City of Duluth based on systematically collected and analyzed data. This presentation will share the results of the survey. This presentation of the Speak Up Speak Out survey data represents support for local decision-making, technical assistance and data analysis. The data were collected and analyzed through advice and consultation with USEPA Office of Research and Development, an

  7. Her Voice Lingers on and Her Memory Is Strategic: Effects of Gender on Directed Forgetting

    PubMed Central

    Yang, Hwajin; Yang, Sujin; Park, Giho

    2013-01-01

    The literature on directed forgetting has employed exclusively visual words. Thus, the potentially interesting aspects of a spoken utterance, which include not only vocal cues (e.g., prosody) but also the speaker and the listener, have been neglected. This study demonstrates that prosody alone does not influence directed-forgetting effects, while the sex of the speaker and the listener significantly modulate directed-forgetting effects for spoken utterances. Specifically, forgetting costs were attenuated for female-spoken items compared to male-spoken items, and forgetting benefits were eliminated among female listeners but not among male listeners. These results suggest that information conveyed in a female voice draws attention to its distinct perceptual attributes, thus interfering with retention of the semantic meaning, while female listeners' superior capacity for processing the surface features of spoken utterances may predispose them to spontaneously employ adaptive strategies to retain content information despite distraction by perceptual features. Our findings underscore the importance of sex differences when processing spoken messages in directed forgetting. PMID:23691141

  8. Predicted singers' vocal fold lengths and voice classification-a study of x-ray morphological measures.

    PubMed

    Roers, Friederike; Mürbe, Dirk; Sundberg, Johan

    2009-07-01

    Students admitted to the solo singing education at the University of Music Dresden, Germany have been submitted to a detailed physical examination of a variety of factors with relevance to voice function since 1959. In the years 1959-1991, this scheme of examinations included X-ray profiles of the singers' vocal tracts. This material of 132 X-rays of voice professionals was used to investigate different laryngeal morphological measures and their relation to vocal fold length. Further, the study aimed to investigate if there are consistent anatomical differences between singers of different voice classifications. The study design used was a retrospective analysis. Vocal fold length could be measured in 29 of these singer subjects directly. These data showed a strong correlation with the anterior-posterior diameter of the subglottis and the trachea as well as with the distance from the anterior contour of the thyroid cartilage to the anterior contour of the spine. These relations were used in an attempt to predict the 132 singers' vocal fold lengths. The results revealed a clear covariation between predicted vocal fold length and voice classification. Anterior-posterior subglottic-tracheal diameter yielded mean vocal fold lengths of 14.9, 16.0, 16.6, 18.4, 19.5, and 20.9mm for sopranos, mezzo-sopranos, altos, tenors, baritones, and basses, respectively. The data support the assumption that there are consistent anatomical laryngeal differences between singers of different voice classifications, which are of relevance to pitch range and timbre of the voice.

  9. Electronic Delivery System: Presentation Features.

    DTIC Science & Technology

    1981-04-01

    THE INFOR’"TiO 1. 0 THE FULNCTIONALITY OF THE PRESENTATIO,’, NOT ITS REPLIC., NATURE IS WHAT COUNTS. S-12 REAL ISM _(CNTD. ) * A SEQUENCE OF...E.G, A MOUSE) IS USED FOR INPUTTINZ RESPONSES, THEY CAN BE VERY EFFICIENT, , S-21 -~i INTERACTION - MECHANISt, S (CONTD.) * TOUCH PANELS -- NATURAL , NO...INTERACTION - MECHANISMS (CONTD, i fm O VOICE INPUT --USED WHERE HANDS OR EYES ARE BUSY (E.G., FOR MAINTENANCE AIDING), -- A NATURAL MEANS OF CO;r UNICATION

  10. A Phonological Rules System

    DTIC Science & Technology

    1975-01-24

    oorrectinq input and a command for entering edit mode with current definitions. 10. 1 THE EDITOR The editor is automatically entered when a sy ...pat-part>::=<consonant- naBe >|l <reduced-name>( <f ull-vowel-naiOf <explici t-stress> 11 <class-naine>|<place- naBe >i <kind- naBe >| VCICE...test>| < voice-test> | (<cond-body>) <kind-test>: : = KIND (EQINQ) fKIND|<)cind- naBe >| <class-test>::=CLASS (BQ|NQ

  11. Enhancing the incorporation of the patient's voice in drug development and evaluation.

    PubMed

    Chalasani, Meghana; Vaidya, Pujita; Mullin, Theresa

    2018-01-01

    People living with a condition are uniquely positioned to inform the understanding of the therapeutic context for drug development and evaluation. In 2012, the U.S. Food and Drug Administration (FDA) established the Patient-Focused Drug Development (PFDD) initiative to more systematically obtain the patient perspective on specific diseases and their currently available treatments. PFDD meetings are unique among FDA public meetings, with a format designed to engage patients and elicit their perspectives on two topic areas: (1) the most significant symptoms of their condition and the impact of the condition on daily life; and, (2) their current approaches to treatment. FDA has conducted 24 disease-specific PFDD meetings to date. The lessons learned from PFDD meetings range from experiences common across rare diseases to more disease specific experiences that matter most to patients. FDA recognizes that FDA-led PFDD meetings alone cannot address the gaps in information on the patient perspective. Patient-focused drug development is an ongoing effort and FDA looks forward to the next steps in advancing the science and the utilization of patient input throughout drug development and evaluation. The U.S. Food and Drug Administration (FDA) has multiple mechanisms for its regulators and staff to interact with patients -- but none quite like its novel Patient-Focused Drug Development (PFDD) initiative. FDA established the PFDD initiative to more systematically obtain the patient perspective on specific diseases and their currently available treatments. Since the initiative's inception in 2012, FDA has held 24 PFDD meetings, covering a range of disease areas and hearing directly from thousands of patients and caregivers. FDA's PFDD meetings have also provided key stakeholders, including patient advocates, researchers, drug developers, healthcare providers, and other government officials, an opportunity to hear the patient's voice. The lessons learned include but are not limited to specific experiences that matter most to patients, patient perspectives on meaningful treatment benefits and how patients want to be engaged in the drug development process. FDA recognizes that FDA-led PFDD meetings alone cannot address the gaps in information on the patient perspective. Further enhancing the incorporation of the patient's voice in drug development and evaluation continues to be a priority for FDA.

  12. Multisensory perception of the six basic emotions is modulated by attentional instruction and unattended modality

    PubMed Central

    Takagi, Sachiko; Hiramatsu, Saori; Tabei, Ken-ichi; Tanaka, Akihiro

    2015-01-01

    Previous studies have shown that the perception of facial and vocal affective expressions interacts with each other. Facial expressions usually dominate vocal expressions when we perceive the emotions of face–voice stimuli. In most of these studies, participants were instructed to pay attention to the face or voice. Few studies compared the perceived emotions with and without specific instructions regarding the modality to which attention should be directed. Also, these studies used combinations of the face and voice which expresses two opposing emotions, which limits the generalizability of the findings. The purpose of this study is to examine whether the emotion perception is modulated by instructions to pay attention to the face or voice using the six basic emotions. Also we examine the modality dominance between the face and voice for each emotion category. Before the experiment, we recorded faces and voices which expresses the six basic emotions and orthogonally combined these faces and voices. Consequently, the emotional valence of visual and auditory information was either congruent or incongruent. In the experiment, there were unisensory and multisensory sessions. The multisensory session was divided into three blocks according to whether an instruction was given to pay attention to a given modality (face attention, voice attention, and no instruction). Participants judged whether the speaker expressed happiness, sadness, anger, fear, disgust, or surprise. Our results revealed that instructions to pay attention to one modality and congruency of the emotions between modalities modulated the modality dominance, and the modality dominance is differed for each emotion category. In particular, the modality dominance for anger changed according to each instruction. Analyses also revealed that the modality dominance suggested by the congruency effect can be explained in terms of the facilitation effect and the interference effect. PMID:25698945

  13. Effects of task performance, helping, voice, and organizational loyalty on performance appraisal ratings.

    PubMed

    Whiting, Steven W; Podsakoff, Philip M; Pierce, Jason R

    2008-01-01

    Despite the fact that several studies have investigated the relationship between organizational citizenship behavior and performance appraisal ratings, the vast majority of these studies have been cross-sectional, correlational investigations conducted in organizational settings that do not allow researchers to establish the causal nature of this relationship. To address this lack of knowledge regarding causality, the authors conducted 2 studies designed to investigate the effects of task performance, helping behavior, voice, and organizational loyalty on performance appraisal evaluations. Findings demonstrated that each of these forms of behavior has significant effects on performance evaluation decisions and suggest that additional attention should be directed at both voice and organizational loyalty as important forms of citizenship behavior aimed at the organization. 2008 APA

  14. Satellite voice broadcase system study. Volume 1: Executive summary

    NASA Technical Reports Server (NTRS)

    Horstein, M.

    1985-01-01

    The feasibility of providing Voice of America (VOA) broadcasts by satellite relay was investigated. Satellite voice broadcast systems are described for three different frequency bands: HF, FHV, and L-band. Geostationary satellite configurations are considered for both frequency bands. A system of subsynchronous, circular satellites with an orbit period of 8 hours was developed for the HF band. The VHF broadcasts are provided by a system of Molniya satellites. The satellite designs are limited in size and weight to the capability of the STS/Centaur launch vehicle combination. At L-band, only four geostationary satellites are needed to meet the requirements of the complete broadcast schedule. These satellites are comparable in size and weight to current satellites designed for the direct broadcast of video program material.

  15. Domestic dogs and puppies can use human voice direction referentially.

    PubMed

    Rossano, Federico; Nitzschner, Marie; Tomasello, Michael

    2014-06-22

    Domestic dogs are particularly skilled at using human visual signals to locate hidden food. This is, to our knowledge, the first series of studies that investigates the ability of dogs to use only auditory communicative acts to locate hidden food. In a first study, from behind a barrier, a human expressed excitement towards a baited box on either the right or left side, while sitting closer to the unbaited box. Dogs were successful in following the human's voice direction and locating the food. In the two following control studies, we excluded the possibility that dogs could locate the box containing food just by relying on smell, and we showed that they would interpret a human's voice direction in a referential manner only when they could locate a possible referent (i.e. one of the boxes) in the environment. Finally, in a fourth study, we tested 8-14-week-old puppies in the main experimental test and found that those with a reasonable amount of human experience performed overall even better than the adult dogs. These results suggest that domestic dogs' skills in comprehending human communication are not based on visual cues alone, but are instead multi-modal and highly flexible. Moreover, the similarity between young and adult dogs' performances has important implications for the domestication hypothesis.

  16. Domestic dogs and puppies can use human voice direction referentially

    PubMed Central

    Rossano, Federico; Nitzschner, Marie; Tomasello, Michael

    2014-01-01

    Domestic dogs are particularly skilled at using human visual signals to locate hidden food. This is, to our knowledge, the first series of studies that investigates the ability of dogs to use only auditory communicative acts to locate hidden food. In a first study, from behind a barrier, a human expressed excitement towards a baited box on either the right or left side, while sitting closer to the unbaited box. Dogs were successful in following the human's voice direction and locating the food. In the two following control studies, we excluded the possibility that dogs could locate the box containing food just by relying on smell, and we showed that they would interpret a human's voice direction in a referential manner only when they could locate a possible referent (i.e. one of the boxes) in the environment. Finally, in a fourth study, we tested 8–14-week-old puppies in the main experimental test and found that those with a reasonable amount of human experience performed overall even better than the adult dogs. These results suggest that domestic dogs’ skills in comprehending human communication are not based on visual cues alone, but are instead multi-modal and highly flexible. Moreover, the similarity between young and adult dogs’ performances has important implications for the domestication hypothesis. PMID:24807249

  17. Spiking and Excitatory/Inhibitory Input Dynamics of Barrel Cells in Response to Whisker Deflections of Varying Velocity and Angular Direction.

    PubMed

    Patel, Mainak

    2018-01-15

    The spiking of barrel regular-spiking (RS) cells is tuned for both whisker deflection direction and velocity. Velocity tuning arises due to thalamocortical (TC) synchrony (but not spike quantity) varying with deflection velocity, coupled with feedforward inhibition, while direction selectivity is not fully understood, though may be due partly to direction tuning of TC spiking. Data show that as deflection direction deviates from the preferred direction of an RS cell, excitatory input to the RS cell diminishes minimally, but temporally shifts to coincide with the time-lagged inhibitory input. This work constructs a realistic large-scale model of a barrel; model RS cells exhibit velocity and direction selectivity due to TC input dynamics, with the experimentally observed sharpening of direction tuning with decreasing velocity. The model puts forth the novel proposal that RS→RS synapses can naturally and simply account for the unexplained direction dependence of RS cell inputs - as deflection direction deviates from the preferred direction of an RS cell, and TC input declines, RS→RS synaptic transmission buffers the decline in total excitatory input and causes a shift in timing of the excitatory input peak from the peak in TC input to the delayed peak in RS input. The model also provides several experimentally testable predictions on the velocity dependence of RS cell inputs. This model is the first, to my knowledge, to study the interaction of direction and velocity and propose physiological mechanisms for the stimulus dependence in the timing and amplitude of RS cell inputs. Copyright © 2017 IBRO. Published by Elsevier Ltd. All rights reserved.

  18. Arizona Telemedicine Program Interprofessional Learning Center: facility design and curriculum development.

    PubMed

    Weinstein, Ronald S; López, Ana Mariá; Barker, Gail P; Krupinski, Elizabeth A; Beinar, Sandra J; Major, Janet; Skinner, Tracy; Holcomb, Michael J; McNeely, Richard A

    2007-10-01

    The Institute for Advanced Telemedicine and Telehealth (i.e., T-Health Institute), a division of the state-wide Arizona Telemedicine Program (ATP), specializes in the creation of innovative health care education programs. This paper describes a first-of-a-kind video amphitheater specifically designed to promote communication within heterogeneous student groups training in the various health care professions. The amphitheater has an audio-video system that facilitates the assembly of ad hoc "in-the-room" electronic interdisciplinary student groups. Off-site faculty members and students can be inserted into groups by video conferencing. When fully implemented, every student will have a personal video camera trained on them, a head phone/microphone, and a personal voice channel. A command and control system will manage the video inputs of the individual participant's head-and-shoulder video images. An audio mixer will manage the separate voice channels of the individual participants and mix them into individual group-specific voice channels for use by the groups' participants. The audio-video system facilitates the easy reconfiguration of the interprofessional electronic groups, viewed on the video wall, without the individual participants in the electronic groups leaving their seats. The amphitheater will serve as a classroom as well as a unique education research laboratory.

  19. Subglottal Impedance-Based Inverse Filtering of Voiced Sounds Using Neck Surface Acceleration

    PubMed Central

    Zañartu, Matías; Ho, Julio C.; Mehta, Daryush D.; Hillman, Robert E.; Wodicka, George R.

    2014-01-01

    A model-based inverse filtering scheme is proposed for an accurate, non-invasive estimation of the aerodynamic source of voiced sounds at the glottis. The approach, referred to as subglottal impedance-based inverse filtering (IBIF), takes as input the signal from a lightweight accelerometer placed on the skin over the extrathoracic trachea and yields estimates of glottal airflow and its time derivative, offering important advantages over traditional methods that deal with the supraglottal vocal tract. The proposed scheme is based on mechano-acoustic impedance representations from a physiologically-based transmission line model and a lumped skin surface representation. A subject-specific calibration protocol is used to account for individual adjustments of subglottal impedance parameters and mechanical properties of the skin. Preliminary results for sustained vowels with various voice qualities show that the subglottal IBIF scheme yields comparable estimates with respect to current aerodynamics-based methods of clinical vocal assessment. A mean absolute error of less than 10% was observed for two glottal airflow measures –maximum flow declination rate and amplitude of the modulation component– that have been associated with the pathophysiology of some common voice disorders caused by faulty and/or abusive patterns of vocal behavior (i.e., vocal hyperfunction). The proposed method further advances the ambulatory assessment of vocal function based on the neck acceleration signal, that previously have been limited to the estimation of phonation duration, loudness, and pitch. Subglottal IBIF is also suitable for other ambulatory applications in speech communication, in which further evaluation is underway. PMID:25400531

  20. How silent is silent reading? Intracerebral evidence for top-down activation of temporal voice areas during reading.

    PubMed

    Perrone-Bertolotti, Marcela; Kujala, Jan; Vidal, Juan R; Hamame, Carlos M; Ossandon, Tomas; Bertrand, Olivier; Minotti, Lorella; Kahane, Philippe; Jerbi, Karim; Lachaux, Jean-Philippe

    2012-12-05

    As you might experience it while reading this sentence, silent reading often involves an imagery speech component: we can hear our own "inner voice" pronouncing words mentally. Recent functional magnetic resonance imaging studies have associated that component with increased metabolic activity in the auditory cortex, including voice-selective areas. It remains to be determined, however, whether this activation arises automatically from early bottom-up visual inputs or whether it depends on late top-down control processes modulated by task demands. To answer this question, we collaborated with four epileptic human patients recorded with intracranial electrodes in the auditory cortex for therapeutic purposes, and measured high-frequency (50-150 Hz) "gamma" activity as a proxy of population level spiking activity. Temporal voice-selective areas (TVAs) were identified with an auditory localizer task and monitored as participants viewed words flashed on screen. We compared neural responses depending on whether words were attended or ignored and found a significant increase of neural activity in response to words, strongly enhanced by attention. In one of the patients, we could record that response at 800 ms in TVAs, but also at 700 ms in the primary auditory cortex and at 300 ms in the ventral occipital temporal cortex. Furthermore, single-trial analysis revealed a considerable jitter between activation peaks in visual and auditory cortices. Altogether, our results demonstrate that the multimodal mental experience of reading is in fact a heterogeneous complex of asynchronous neural responses, and that auditory and visual modalities often process distinct temporal frames of our environment at the same time.

  1. The effect of deep brain stimulation on the speech motor system.

    PubMed

    Mücke, Doris; Becker, Johannes; Barbe, Michael T; Meister, Ingo; Liebhart, Lena; Roettger, Timo B; Dembek, Till; Timmermann, Lars; Grice, Martine

    2014-08-01

    Chronic deep brain stimulation of the nucleus ventralis intermedius is an effective treatment for individuals with medication-resistant essential tremor. However, these individuals report that stimulation has a deleterious effect on their speech. The present study investigates one important factor leading to these effects: the coordination of oral and glottal articulation. Sixteen native-speaking German adults with essential tremor, between 26 and 86 years old, with and without chronic deep brain stimulation of the nucleus ventralis intermedius and 12 healthy, age-matched subjects were recorded performing a fast syllable repetition task (/papapa/, /tatata/, /kakaka/). Syllable duration and voicing-to-syllable ratio as well as parameters related directly to consonant production, voicing during constriction, and frication during constriction were measured. Voicing during constriction was greater in subjects with essential tremor than in controls, indicating a perseveration of voicing into the voiceless consonant. Stimulation led to fewer voiceless intervals (voicing-to-syllable ratio), indicating a reduced degree of glottal abduction during the entire syllable cycle. Stimulation also induced incomplete oral closures (frication during constriction), indicating imprecise oral articulation. The detrimental effect of stimulation on the speech motor system can be quantified using acoustic measures at the subsyllabic level.

  2. The management of vocal fold nodules in children: a national survey of speech-language pathologists.

    PubMed

    Signorelli, Monique E; Madill, Catherine J; McCabe, Patricia

    2011-06-01

    The purpose of this study was to determine the management options and voice therapy techniques currently being used by practicing speech-language pathologists (SLPs) to treat vocal fold nodules (VFNs) in children. The sources used by SLPs to inform and guide their clinical decisions when managing VFNs in children were also explored. Sixty-two SLPs completed a 23-item web-based survey. Data was analysed using frequency counts, content analyses, and supplementary analyses. SLPs reported using a range of management options and voice therapy techniques to treat VFNs in children. Voice therapy was reportedly the most frequently used management option across all respondents, with the majority of SLPs using a combination of indirect and direct voice therapy techniques. When selecting voice therapy techniques, the majority of SLPs reported that they did not use the limited external evidence available to guide their clinical decisions. Additionally, the majority of SLPs reported that they frequently relied on lower levels of evidence or non-evidence-based sources to guide clinical practice both in the presence and absence of higher quality evidence. Further research needs to investigate strategies to remove the barriers that impede SLPs use of external evidence when managing VFNs in children.

  3. Modeling and Analysis of Hybrid Cellular/WLAN Systems with Integrated Service-Based Vertical Handoff Schemes

    NASA Astrophysics Data System (ADS)

    Xia, Weiwei; Shen, Lianfeng

    We propose two vertical handoff schemes for cellular network and wireless local area network (WLAN) integration: integrated service-based handoff (ISH) and integrated service-based handoff with queue capabilities (ISHQ). Compared with existing handoff schemes in integrated cellular/WLAN networks, the proposed schemes consider a more comprehensive set of system characteristics such as different features of voice and data services, dynamic information about the admitted calls, user mobility and vertical handoffs in two directions. The code division multiple access (CDMA) cellular network and IEEE 802.11e WLAN are taken into account in the proposed schemes. We model the integrated networks by using multi-dimensional Markov chains and the major performance measures are derived for voice and data services. The important system parameters such as thresholds to prioritize handoff voice calls and queue sizes are optimized. Numerical results demonstrate that the proposed ISHQ scheme can maximize the utilization of overall bandwidth resources with the best quality of service (QoS) provisioning for voice and data services.

  4. Voice-Dictated versus Typed-in Clinician Notes: Linguistic Properties and the Potential Implications on Natural Language Processing

    PubMed Central

    Zheng, Kai; Mei, Qiaozhu; Yang, Lei; Manion, Frank J.; Balis, Ulysses J.; Hanauer, David A.

    2011-01-01

    In this study, we comparatively examined the linguistic properties of narrative clinician notes created through voice dictation versus those directly entered by clinicians via a computer keyboard. Intuitively, the nature of voice-dictated notes would resemble that of natural language, while typed-in notes may demonstrate distinctive language features for reasons such as intensive usage of acronyms. The study analyses were based on an empirical dataset retrieved from our institutional electronic health records system. The dataset contains 30,000 voice-dictated notes and 30,000 notes that were entered manually; both were encounter notes generated in ambulatory care settings. The results suggest that between the narrative clinician notes created via these two different methods, there exists a considerable amount of lexical and distributional differences. Such differences could have a significant impact on the performance of natural language processing tools, necessitating these two different types of documents being differentially treated. PMID:22195229

  5. Relationship between perceived politeness and spectral characteristics of voice

    NASA Astrophysics Data System (ADS)

    Ito, Mika

    2005-04-01

    This study investigates the role of voice quality in perceiving politeness under conditions of varying relative social status among Japanese male speakers. The work focuses on four important methodological issues: experimental control of sociolinguistic aspects, eliciting natural spontaneous speech, obtaining recording quality suitable for voice quality analysis, and assessment of glottal characteristics through the use of non-invasive direct measurements of the speech spectrum. To obtain natural, unscripted utterances, the speech data were collected with a Map Task. This methodology allowed us to study the effect of manipulating relative social status among participants in the same community. We then computed the relative amplitudes of harmonics and formant peaks in spectra obtained from the Map Task recordings. Finally, an experiment was conducted to observe the alignment between acoustic measures and the perceived politeness of the voice samples. The results suggest that listeners' perceptions of politeness are determined by spectral characteristics of speakers, in particular, spectral tilts obtained by computing the difference in amplitude between the first harmonic and the third formant.

  6. Henry's voices: the representation of auditory verbal hallucinations in an autobiographical narrative.

    PubMed

    Demjén, Zsófia; Semino, Elena

    2015-06-01

    The book Henry's Demons (2011) recounts the events surrounding Henry Cockburn's diagnosis of schizophrenia from the alternating perspectives of Henry himself and his father Patrick. In this paper, we present a detailed linguistic analysis of Henry's first-person accounts of experiences that could be described as auditory verbal hallucinations. We first provide a typology of Henry's voices, taking into account who or what is presented as speaking, what kinds of utterances they produce and any salient stylistic features of these utterances. We then discuss the linguistically distinctive ways in which Henry represents these voices in his narrative. We focus on the use of Direct Speech as opposed to other forms of speech presentation, the use of the sensory verbs hear and feel and the use of 'non-factive' expressions such as I thought and as if. We show how different linguistic representations may suggest phenomenological differences between the experience of hallucinatory voices and the perception of voices that other people can also hear. We, therefore, propose that linguistic analysis is ideally placed to provide in-depth accounts of the phenomenology of voice hearing and point out the implications of this approach for clinical practice and mental healthcare. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

  7. The effect of voice quality and competing speakers in a passage comprehension task: perceived effort in relation to cognitive functioning and performance in children with normal hearing.

    PubMed

    von Lochow, Heike; Lyberg-Åhlander, Viveka; Sahlén, Birgitta; Kastberg, Tobias; Brännström, K Jonas

    2018-04-01

    The study investigates the effect of voice quality and competing speakers on perceived effort in a passage comprehension task in relation to cognitive functioning. In addition, it explores if perceived effort was related to performance. A total of 49 children (aged 7:03 to 12:02 years) with normal hearing participated. The children performed an auditory passage comprehension task presented with six different listening conditions consisting of a typical voice or a dysphonic voice presented in quiet, with one competing speaker, and with four competing speakers. After completing the task, they rated their perceived effort on a five-grade scale. The children also performed tasks measuring working memory capacity (WMC) and executive functioning. The results show that voice quality had no direct effect on perceived effort but the children's ratings of perceived effort were related to their executive functioning. A significant effect was seen for background listening condition indicating higher perceived effort for background listening conditions with competing speakers. The effects of background listening condition were mainly related to the children's WMC but also their executive functioning. It can be concluded that the individual susceptibility to the effect of the dysphonic voice is related to the child's executive functioning. The individual susceptibility to the presence of competing speakers is related to the child's WMC and executive functioning.

  8. Functional hoarseness in children: short-term play therapy with family dynamic counseling as therapy of choice.

    PubMed

    Kollbrunner, Jürg; Seifert, Eberhard

    2013-09-01

    Children with nonorganic voice disorders (NVDs) are treated mainly using direct voice therapy techniques such as the accent method or glottal attack changes and indirect methods such as vocal hygiene and voice education. However, both approaches tackle only the symptoms and not etiological factors in the family dynamics and therefore often enjoy little success. The aim of the "Bernese Brief Dynamic Intervention" (BBDI) for children with NVD was to extend the effectiveness of pediatric voice therapies with a psychosomatic concept combining short-term play therapy with the child and family dynamic counseling of the parents. This study compares the therapeutic changes in three groups where different procedures were used, before intervention and 1 year afterward: counseling of parents (one to two consultations; n = 24), Brief Dynamic Intervention on the lines of the BBDI (three to five play therapy sessions with the child plus two to four sessions with the parents; n = 20), and traditional voice therapy (n = 22). A Voice Questionnaire for Parents developed by us with 59 questions to be answered on a four-point Likert scale was used to measure the change. According to the parents' assessment, a significant improvement in voice quality was achieved in all three methods. Counseling of parents (A) appears to have led parents to give their child more latitude, for example, they stopped nagging the child or demanding that he/she should behave strictly by the rules. After BBDI (B), the mothers were more responsive to their children's wishes and the children were more relaxed and their speech became livelier. At home, they called out to them less often at a distance, which probably improved parent-child dialog. Traditional voice therapy (C) seems to have had a positive effect on the children's social competence. BBDI seems to have the deepest, widest, and therefore probably the most enduring therapeutic effect on children with NVD. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  9. Template Based Low Data Rate Speech Encoder

    DTIC Science & Technology

    1993-09-30

    Nasality Distinguishes In/ from d/ 95.6 96.9 1m/ from /b/, etc. Sustention Distinguishes /f/ from /p/, $7.5 88.3 ibi from N/, Al from /0 8. etc. Sibilation...processor performs mainly Processor Workstation input/output (I/O) operations. The dynamic random access memory (DRAM) has 16 million bytes of...storage capacity. To execute the 800-b/s voice algorithm, the following amount of memory is needed: 5 MB for tables, 1.5 MB for it "program, and 30 KB for

  10. Voice Recognition as an Input Modality for the TACCO Preflight Data Insertion Task in the P-3C Aircraft

    DTIC Science & Technology

    1981-03-01

    C., the9nr aooearei as a Ii kel y candida-- tte for thin simulationi crcaram lanauage for a number of reasons: 1. Tt is a structured lanquaaie with...taonl’eus to j-jv fil s1 ~ lto e I INDEY 1Iidisolav TN~ ala n- u h iuain * isnlIayI IL’FX tte au 0&-d "L IJLtpopJ atrML si ’njlI t ion terotinat ion i s

  11. Voice Signals Produced With Jitter Through a Stochastic One-mass Mechanical Model.

    PubMed

    Cataldo, Edson; Soize, Christian

    2017-01-01

    The quasiperiodic oscillation of the vocal folds causes perturbations in the length of the glottal cycles, which are known as jitter. The observation of the glottal cycles variations suggests that jitter is a random phenomenon described by random deviations of the glottal cycle lengths in relation to a corresponding mean value and, in general, its values are expressed as a percentage of the duration of the glottal pulse. The objective of this paper is the construction of a stochastic model for jitter using a one-mass mechanical model of the vocal folds, which assumes complete right-left symmetry of the vocal folds, and which considers motions of the vocal folds only in the horizontal direction. The jitter has been the subject for researchers due to its important applications such as the identification of pathological voices (nodules in the vocal folds, paralysis of the vocal folds, or even, the vocal aging, among others). Large values for jitter variations can indicate a pathological characteristic of the voice. The corresponding stiffness of each vocal fold is considered as a stochastic process, and its modeling is proposed. The probability density function of the fundamental frequency related to the voice signals produced are constructed and compared for different levels of jitter. Some samples of synthesized voices in these cases are obtained. It is showed that jitter could be obtained using the model proposed. The Praat software was also used to verify the measures of jitter in the synthesized voice signals. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  12. A Low-Cost, Man-Portable, Free-Space Optics Communications Device for Ethernet Applications

    DTIC Science & Technology

    2004-12-01

    ix LIST OF FIGURES Figure 1. Patent for the Photophone filed by Alexander Graham Bell and Charles S. Tainter. (From Ref. [8...Tainter patented a device they called the Photophone in 1880 (Fig. 1.) By using a series of mirrors and lenses, they were able to modulate a voice...signal on to a ray of sunlight and send it to a receiver 200 meters away [8]. In the Photophone , voice sound waves were directed on to a mirror that

  13. DigitalVHI--a freeware open-source software application to capture the Voice Handicap Index and other questionnaire data in various languages.

    PubMed

    Herbst, Christian T; Oh, Jinook; Vydrová, Jitka; Švec, Jan G

    2015-07-01

    In this short report we introduce DigitalVHI, a free open-source software application for obtaining Voice Handicap Index (VHI) and other questionnaire data, which can be put on a computer in clinics and used in clinical practice. The software can simplify performing clinical studies since it makes the VHI scores directly available for analysis in a digital form. It can be downloaded from http://www.christian-herbst.org/DigitalVHI/.

  14. Home Diabetes Monitoring through Touch-Tone Computer Data Entry and Voice Synthesizer Response

    PubMed Central

    Arbogast, James G.; Dodrill, William H.

    1984-01-01

    Current studies suggest that the control of Diabetes mellitus can be improved with home monitoring of blood sugars. Voice synthesizers and recent technology, allowing decoding of Touch-Tone® pulses into their digital equivalents, make it possible for diabetics with no more sophisticated equipment than a Touch-Tone® telephone to enter their blood sugars directly into a medical office computer. A working prototype that can provide physicians with timely, logically oriented information about their diabetics is discussed along with plans to expand this concept into giving the patients uncomplicated therapeutic advice without the need for a direct patient/physician interaction. The potential impact on health care costs and the management of other chronic diseases is presented.

  15. Comparing the experience of voices in borderline personality disorder with the experience of voices in a psychotic disorder: A systematic review.

    PubMed

    Merrett, Zalie; Rossell, Susan L; Castle, David J

    2016-07-01

    In clinical settings, there is substantial evidence both clinically and empirically to suggest that approximately 50% of individuals with borderline personality disorder experience auditory verbal hallucinations. However, there is limited research investigating the phenomenology of these voices. The aim of this study was to review and compare our current understanding of auditory verbal hallucinations in borderline personality disorder with auditory verbal hallucinations in patients with a psychotic disorder, to critically analyse existing studies investigating auditory verbal hallucinations in borderline personality disorder and to identify gaps in current knowledge, which will help direct future research. The literature was searched using the electronic database Scopus, PubMed and MEDLINE. Relevant studies were included if they were written in English, were empirical studies specifically addressing auditory verbal hallucinations and borderline personality disorder, were peer reviewed, used only adult humans and sample comprising borderline personality disorder as the primary diagnosis, and included a comparison group with a primary psychotic disorder such as schizophrenia. Our search strategy revealed a total of 16 articles investigating the phenomenology of auditory verbal hallucinations in borderline personality disorder. Some studies provided evidence to suggest that the voice experiences in borderline personality disorder are similar to those experienced by people with schizophrenia, for example, occur inside the head, and often involved persecutory voices. Other studies revealed some differences between schizophrenia and borderline personality disorder voice experiences, with the borderline personality disorder voices sounding more derogatory and self-critical in nature and the voice-hearers' response to the voices were more emotionally resistive. Furthermore, in one study, the schizophrenia group's voices resulted in more disruption in daily functioning. These studies are, however, limited in number and do not provide definitive evidence of these differences. The limited research examining auditory verbal hallucinations experiences in borderline personality disorder poses a significant diagnostic and treatment challenge. A deeper understanding of the precise phenomenological characteristics will help us in terms of diagnostic distinction as well as inform treatments. © The Royal Australian and New Zealand College of Psychiatrists 2016.

  16. Detecting "Infant-Directedness" in Face and Voice

    ERIC Educational Resources Information Center

    Kim, Hojin I.; Johnson, Scott P.

    2014-01-01

    Five- and 3-month-old infants' perception of infant-directed (ID) faces and the role of speech in perceiving faces were examined. Infants' eye movements were recorded as they viewed a series of two side-by-side talking faces, one infant-directed and one adult-directed (AD), while listening to ID speech, AD speech, or in silence. Infants…

  17. Domain-specific impairment of source memory following a right posterior medial temporal lobe lesion.

    PubMed

    Peters, Jan; Koch, Benno; Schwarz, Michael; Daum, Irene

    2007-01-01

    This single case analysis of memory performance in a patient with an ischemic lesion affecting posterior but not anterior right medial temporal lobe (MTL) indicates that source memory can be disrupted in a domain-specific manner. The patient showed normal recognition memory for gray-scale photos of objects (visual condition) and spoken words (auditory condition). While memory for visual source (texture/color of the background against which pictures appeared) was within the normal range, auditory source memory (male/female speaker voice) was at chance level, a performance pattern significantly different from the control group. This dissociation is consistent with recent fMRI evidence of anterior/posterior MTL dissociations depending upon the nature of source information (visual texture/color vs. auditory speaker voice). The findings are in good agreement with the view of dissociable memory processing by the perirhinal cortex (anterior MTL) and parahippocampal cortex (posterior MTL), depending upon the neocortical input that these regions receive. (c) 2007 Wiley-Liss, Inc.

  18. Automatic measurement of voice onset time using discriminative structured prediction.

    PubMed

    Sonderegger, Morgan; Keshet, Joseph

    2012-12-01

    A discriminative large-margin algorithm for automatic measurement of voice onset time (VOT) is described, considered as a case of predicting structured output from speech. Manually labeled data are used to train a function that takes as input a speech segment of an arbitrary length containing a voiceless stop, and outputs its VOT. The function is explicitly trained to minimize the difference between predicted and manually measured VOT; it operates on a set of acoustic feature functions designed based on spectral and temporal cues used by human VOT annotators. The algorithm is applied to initial voiceless stops from four corpora, representing different types of speech. Using several evaluation methods, the algorithm's performance is near human intertranscriber reliability, and compares favorably with previous work. Furthermore, the algorithm's performance is minimally affected by training and testing on different corpora, and remains essentially constant as the amount of training data is reduced to 50-250 manually labeled examples, demonstrating the method's practical applicability to new datasets.

  19. High-frequency energy in singing and speech

    NASA Astrophysics Data System (ADS)

    Monson, Brian Bruce

    While human speech and the human voice generate acoustical energy up to (and beyond) 20 kHz, the energy above approximately 5 kHz has been largely neglected. Evidence is accruing that this high-frequency energy contains perceptual information relevant to speech and voice, including percepts of quality, localization, and intelligibility. The present research was an initial step in the long-range goal of characterizing high-frequency energy in singing voice and speech, with particular regard for its perceptual role and its potential for modification during voice and speech production. In this study, a database of high-fidelity recordings of talkers was created and used for a broad acoustical analysis and general characterization of high-frequency energy, as well as specific characterization of phoneme category, voice and speech intensity level, and mode of production (speech versus singing) by high-frequency energy content. Directionality of radiation of high-frequency energy from the mouth was also examined. The recordings were used for perceptual experiments wherein listeners were asked to discriminate between speech and voice samples that differed only in high-frequency energy content. Listeners were also subjected to gender discrimination tasks, mode-of-production discrimination tasks, and transcription tasks with samples of speech and singing that contained only high-frequency content. The combination of these experiments has revealed that (1) human listeners are able to detect very subtle level changes in high-frequency energy, and (2) human listeners are able to extract significant perceptual information from high-frequency energy.

  20. Comparison of voice-use profiles between elementary classroom and music teachers.

    PubMed

    Morrow, Sharon L; Connor, Nadine P

    2011-05-01

    Among teachers, music teachers are roughly four times more likely than classroom teachers to develop voice-related problems. Although it has been established that music teachers use their voices at high intensities and durations in the course of their workday, voice-use profiles concerning the amount and intensity of vocal use and vocal load have neither been quantified nor has vocal load for music teachers been compared with classroom teachers using these same voice-use parameters. In this study, total phonation time, fundamental frequency (F₀), and vocal intensity (dB SPL [sound pressure level]) were measured or estimated directly using a KayPENTAX Ambulatory Phonation Monitor (KayPENTAX, Lincoln Park, NJ). Vocal load was calculated as cycle and distance dose, as defined by Švec et al (2003), which integrates total phonation time, F₀, and vocal intensity. Twelve participants (n = 7 elementary music teachers and n = 5 elementary classroom teachers) were monitored during five full teaching days of one workweek to determine average vocal load for these two groups of teachers. Statistically significant differences in all measures were found between the two groups (P < 0.05) with large effect sizes for all parameters. These results suggest that typical vocal loads for music teachers are substantially higher than those experienced by classroom teachers (P < 0.01). This study suggests that reducing vocal load may have immediate clinical and educational benefits in vocal health in music teachers. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  1. Electromyographic activity of strap and cricothyroid muscles in pitch change.

    PubMed

    Roubeau, B; Chevrie-Muller, C; Lacau Saint Guily, J

    1997-05-01

    The EMG activity of the cricothyroid muscle (CT) and the three extrinsic laryngeal muscles (thyohyoid, TH; sternothyroid, ST, and sternohyoid, SH) were recorded throughout the voice range of one female and one male subject, both untrained singers. The voice range was examined using rising and falling glissandos (production of a sustained sound with progressive and continuous variation of fundamental frequency). Muscle activity was observed at various pitches during the glissandos. The strap muscle activity during the production of glissandos appears to be synergistic. At the lowest frequency, the CT is inactive but strap muscles (TH, ST, SH) are active. As frequency increases, strap muscle activity decreases while the CT controls frequency in the middle of the range. At higher frequencies the strap muscles once again become active. This activity might depend on the vocal vibratory mechanism involved. The role of the strap muscles at high pitches is a widely debated point but it seems that in some way they control the phenomena relevant to the rising pitch. The phasic-type strap muscle activity contrasts with the tonic-type activity of the CT. The CT closely controls the frequency, while the straps are not directly linked to the pitch but rather to the evolution of the frequency of voice production (speaking voice, singing voice, held notes, glissandos, trillo, vibrato, etc.).

  2. Developmental Changes in Locating Voice and Sound in Space

    PubMed Central

    Kezuka, Emiko; Amano, Sachiko; Reddy, Vasudevi

    2017-01-01

    We know little about how infants locate voice and sound in a complex multi-modal space. Using a naturalistic laboratory experiment the present study tested 35 infants at 3 ages: 4 months (15 infants), 5 months (12 infants), and 7 months (8 infants). While they were engaged frontally with one experimenter, infants were presented with (a) a second experimenter’s voice and (b) castanet sounds from three different locations (left, right, and behind). There were clear increases with age in the successful localization of sounds from all directions, and a decrease in the number of repetitions required for success. Nonetheless even at 4 months two-thirds of the infants attempted to search for the voice or sound. At all ages localizing sounds from behind was more difficult and was clearly present only at 7 months. Perseverative errors (looking at the last location) were present at all ages and appeared to be task specific (only present in the 7 month-olds for the behind location). Spontaneous attention shifts by the infants between the two experimenters, evident at 7 months, suggest early evidence for infant initiation of triadic attentional engagements. There was no advantage found for voice over castanet sounds in this study. Auditory localization is a complex and contextual process emerging gradually in the first half of the first year. PMID:28979220

  3. The economic impact of vocal attrition in public school teachers in Miami-Dade County.

    PubMed

    Rosow, David E; Szczupak, Mikhaylo; Saint-Victor, Sandra; Gerhard, Julia D; DuPont, Carl; Lo, Kaming

    2016-03-01

    Teachers are a known at-risk population for voice disorders. The prevalence and risk factors for voice disorders have been well studied in this population, but little is known about the associated economic cost. The purpose of this study is to assess the economic impact of voice dysfunction in teachers and understand the difference between the cost of absenteeism and presenteeism as a direct result of voice dysfunction. Cross-sectional analysis via self-administered online questionnaire. A total of 14,256 public school teachers from Miami-Dade County, Florida, were asked to participate. Questions were formatted based on the previously validated Work Productivity and Activity Impairment: Specific Health Problem questionnaire adapted for hoarseness and voice disorders. Additional demographic questions were included in the questionnaire. A total of 961 questionnaire responses were received. The demographic characteristics of respondents closely matched known statistics for public school teachers in Miami-Dade County. Economic calculations were performed for each questionnaire respondent and summed for all respondents to avoid bias. Per week, absenteeism-related costs were $25,000, whereas presenteeism-related costs were approximately $300,000. These figures were used to extrapolate annual cost. Per year, absenteeism-related costs were $1 million, whereas presenteeism-related costs were approximately $12 million. The economic impact of voice dysfunction on the teaching profession is enormous. With the above calculations only including lost wages and decreased productivity, the actual figures may in fact be larger (cost of substitute teachers, impact on nonwork activities, etc.). Research investigating preventative measures for voice dysfunction in teachers is necessary to reduce this costly issue. 2C. Laryngoscope, 126:665-671, 2016. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.

  4. Uncertainty quantification of voice signal production mechanical model and experimental updating

    NASA Astrophysics Data System (ADS)

    Cataldo, E.; Soize, C.; Sampaio, R.

    2013-11-01

    The aim of this paper is to analyze the uncertainty quantification in a voice production mechanical model and update the probability density function corresponding to the tension parameter using the Bayes method and experimental data. Three parameters are considered uncertain in the voice production mechanical model used: the tension parameter, the neutral glottal area and the subglottal pressure. The tension parameter of the vocal folds is mainly responsible for the changing of the fundamental frequency of a voice signal, generated by a mechanical/mathematical model for producing voiced sounds. The three uncertain parameters are modeled by random variables. The probability density function related to the tension parameter is considered uniform and the probability density functions related to the neutral glottal area and the subglottal pressure are constructed using the Maximum Entropy Principle. The output of the stochastic computational model is the random voice signal and the Monte Carlo method is used to solve the stochastic equations allowing realizations of the random voice signals to be generated. For each realization of the random voice signal, the corresponding realization of the random fundamental frequency is calculated and the prior pdf of this random fundamental frequency is then estimated. Experimental data are available for the fundamental frequency and the posterior probability density function of the random tension parameter is then estimated using the Bayes method. In addition, an application is performed considering a case with a pathology in the vocal folds. The strategy developed here is important mainly due to two things. The first one is related to the possibility of updating the probability density function of a parameter, the tension parameter of the vocal folds, which cannot be measured direct and the second one is related to the construction of the likelihood function. In general, it is predefined using the known pdf. Here, it is constructed in a new and different manner, using the own system considered.

  5. Writing about rape: use of the passive voice and other distancing text features as an expression of perceived responsibility of the victim.

    PubMed

    Bohner, G

    2001-12-01

    The hypothesis that the passive voice is used to put the actor in the background and the acted-upon person in the focus of discourse is tested in the realm of sexual violence. German university students (N = 67) watched a silent video segment depicting a rape whose circumstances, depending on condition, could or could not be easily interpreted in terms of rape myths. Then they wrote down what they had seen, judged the responsibility of assailant and victim, and completed a rape-myth acceptance scale. Participants used the passive voice more frequently to describe the rape itself vs. other actions they had watched. When circumstances of the rape were easily interpretable in terms of rape myths, use of the passive voice correlated positively with rape-myth acceptance and perceived responsibility of the victim, and negatively with perceived responsibility of the assailant. The language of headlines that participants generated for their reports also reflected judgments of assailant and victim responsibility. Implications for the non-reactive assessment of responsibility attributions and directions for future research are discussed.

  6. Overview of the Anik C satellites and services

    NASA Astrophysics Data System (ADS)

    Smart, F. H.

    An overview of the important technical characteristics of the Anik C series of Canadian communications satellites is presented. The system was launched as part of the Telesat Communications payload of the Space Shuttle in 1982. Among the services the system will in the near future provide are: a 27 MHz channel bandwidth television service for pay-TV distribution in Canada; two TV channels for hockey broadcasts and a transportable TV system; a heavy-voice route telephone service for five major Canadian cities; and a telephone system for business voice and data communications. Services anticipated for Anik-C satellites later in the decade include a Single Channel Per Carrier (SCPC) voice and data communications system for British Columbia and the Maritime Provinces, and a direct-to-home broadcast service to be sold to television markets in the United States.

  7. Assessing the effectiveness of botulinum toxin injections for adductor spasmodic dysphonia: clinician and patient perception.

    PubMed

    Braden, Maia N; Johns, Michael M; Klein, Adam M; Delgaudio, John M; Gilman, Marina; Hapner, Edie R

    2010-03-01

    To determine the effectiveness of Botox treatment for adductor spasmodic dysphonia (ADSD), the clinician and patient judge changes in voice symptoms and the effect on quality of life. Currently, there is no standard protocol for determining the effectiveness of Botox injections in treating ADSD. Therefore, clinicians use a variety of perceptual scales and patient-based self-assessments to determine patients' impressions of severity and changes after treatments. The purpose of this study was to assess clinician-patient agreement of the effects of Botox on voice quality and quality of life in ADSD. Retrospective chart review of 199 randomly selected patients since 2004. Results indicated a weak correlation between the patient's assessment of voice impairment (EIS) and patient's quality of life impairment (Voice-Related Quality of Life [V-RQOL]) in the mild-moderate dysphonia severity group and the moderate-to-severe dysphonia group. There was a weak correlation between the patient's assessment of voice impairment EIS and the clinician's perceptual judgment of voice impairment (Consensus Auditory Perceptual Evaluation of Voice [CAPE-V]) only in the moderate to severe dysphonia group. There was a weak correlation between the patient's quality of life impairment (V-RQOL) and the clinician's perceptual judgment of voice impairment (CAPE-V) only in the severe to profound dysphonia group. The poor relationship among commonly used outcome measures leads us to question how best to assess the effectiveness of Botox in ADSD. Clinicians are required to document treatment outcomes, making it important to use scales that are valid, reliable, and sensitive to change. Future research directions include examining relationships between measures both before and after Botox injections, examining the specific factors that determine quality of life changes, and further research on specific parameters of the CAPE-V as well as comparing perceptual and quality of life scales with acoustic and aerodynamic measures in this population would be beneficial in the move toward more effective ways of measuring change. Copyright (c) 2010 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  8. Piglets Learn to Use Combined Human-Given Visual and Auditory Signals to Find a Hidden Reward in an Object Choice Task

    PubMed Central

    Bensoussan, Sandy; Cornil, Maude; Meunier-Salaün, Marie-Christine; Tallet, Céline

    2016-01-01

    Although animals rarely use only one sense to communicate, few studies have investigated the use of combinations of different signals between animals and humans. This study assessed for the first time the spontaneous reactions of piglets to human pointing gestures and voice in an object-choice task with a reward. Piglets (Sus scrofa domestica) mainly use auditory signals–individually or in combination with other signals—to communicate with their conspecifics. Their wide hearing range (42 Hz to 40.5 kHz) fits the range of human vocalisations (40 Hz to 1.5 kHz), which may induce sensitivity to the human voice. However, only their ability to use visual signals from humans, especially pointing gestures, has been assessed to date. The current study investigated the effects of signal type (visual, auditory and combined visual and auditory) and piglet experience on the piglets’ ability to locate a hidden food reward over successive tests. Piglets did not find the hidden reward at first presentation, regardless of the signal type given. However, they subsequently learned to use a combination of auditory and visual signals (human voice and static or dynamic pointing gestures) to successfully locate the reward in later tests. This learning process may result either from repeated presentations of the combination of static gestures and auditory signals over successive tests, or from transitioning from static to dynamic pointing gestures, again over successive tests. Furthermore, piglets increased their chance of locating the reward either if they did not go straight to a bowl after entering the test area or if they stared at the experimenter before visiting it. Piglets were not able to use the voice direction alone, indicating that a combination of signals (pointing and voice direction) is necessary. Improving our communication with animals requires adapting to their individual sensitivity to human-given signals. PMID:27792731

  9. Piglets Learn to Use Combined Human-Given Visual and Auditory Signals to Find a Hidden Reward in an Object Choice Task.

    PubMed

    Bensoussan, Sandy; Cornil, Maude; Meunier-Salaün, Marie-Christine; Tallet, Céline

    2016-01-01

    Although animals rarely use only one sense to communicate, few studies have investigated the use of combinations of different signals between animals and humans. This study assessed for the first time the spontaneous reactions of piglets to human pointing gestures and voice in an object-choice task with a reward. Piglets (Sus scrofa domestica) mainly use auditory signals-individually or in combination with other signals-to communicate with their conspecifics. Their wide hearing range (42 Hz to 40.5 kHz) fits the range of human vocalisations (40 Hz to 1.5 kHz), which may induce sensitivity to the human voice. However, only their ability to use visual signals from humans, especially pointing gestures, has been assessed to date. The current study investigated the effects of signal type (visual, auditory and combined visual and auditory) and piglet experience on the piglets' ability to locate a hidden food reward over successive tests. Piglets did not find the hidden reward at first presentation, regardless of the signal type given. However, they subsequently learned to use a combination of auditory and visual signals (human voice and static or dynamic pointing gestures) to successfully locate the reward in later tests. This learning process may result either from repeated presentations of the combination of static gestures and auditory signals over successive tests, or from transitioning from static to dynamic pointing gestures, again over successive tests. Furthermore, piglets increased their chance of locating the reward either if they did not go straight to a bowl after entering the test area or if they stared at the experimenter before visiting it. Piglets were not able to use the voice direction alone, indicating that a combination of signals (pointing and voice direction) is necessary. Improving our communication with animals requires adapting to their individual sensitivity to human-given signals.

  10. Economic Evaluation of Voice Recognition (VR) for the Clinician’s Desktop at the Naval Hospital Roosevelt Roads

    DTIC Science & Technology

    1997-09-01

    first PC-based, very large vocabulary dictation system with a continuous natural language free flow approach to speech recognition. (This system allows...indicating the likelihood that a particular stored HMM reference model is the best match for the input. This approach is called the Baum-Welch...InfoCentral, and Envoy 1.0; and Lotus Development Corp.’s SmartSuite 3, Approach 3.0, and Organizer. 2. IBM At a press conference in New York in June 1997, IBM

  11. Man-machine interfaces in health care

    NASA Technical Reports Server (NTRS)

    Charles, Steve; Williams, Roy E.

    1991-01-01

    The surgeon, like the pilot, is confronted with an ever increasing volume of voice, data, and image input. Simultaneously, the surgeon must control a rapidly growing number of devices to deliver care to the patient. The broad disciplines of man-machine interface design, systems integration, and teleoperation will play a role in the operating room of the future. The purpose of this communication is to report the incorporation of these design concepts into new surgical and laser delivery systems. A review of each general problem area and the systems under development to solve the problems are presented.

  12. Developing Ethical Direction

    ERIC Educational Resources Information Center

    Ribble, Mike S.; Bailey,Gerald D.

    2005-01-01

    When you read or hear an unethical suggestion, such as "Steal this article and sell it to another magazine," we're guessing that your internal compass indicates "wrong direction." In other words, your internal voice says, "No, that would be wrong!" Your internal compass tells you when something is right and something is wrong. In our example, your…

  13. The expert surgical assistant. An intelligent virtual environment with multimodal input.

    PubMed

    Billinghurst, M; Savage, J; Oppenheimer, P; Edmond, C

    1996-01-01

    Virtual Reality has made computer interfaces more intuitive but not more intelligent. This paper shows how an expert system can be coupled with multimodal input in a virtual environment to provide an intelligent simulation tool or surgical assistant. This is accomplished in three steps. First, voice and gestural input is interpreted and represented in a common semantic form. Second, a rule-based expert system is used to infer context and user actions from this semantic representation. Finally, the inferred user actions are matched against steps in a surgical procedure to monitor the user's progress and provide automatic feedback. In addition, the system can respond immediately to multimodal commands for navigational assistance and/or identification of critical anatomical structures. To show how these methods are used we present a prototype sinus surgery interface. The approach described here may easily be extended to a wide variety of medical and non-medical training applications by making simple changes to the expert system database and virtual environment models. Successful implementation of an expert system in both simulated and real surgery has enormous potential for the surgeon both in training and clinical practice.

  14. A SOUND SOURCE LOCALIZATION TECHNIQUE TO SUPPORT SEARCH AND RESCUE IN LOUD NOISE ENVIRONMENTS

    NASA Astrophysics Data System (ADS)

    Yoshinaga, Hiroshi; Mizutani, Koichi; Wakatsuki, Naoto

    At some sites of earthquakes and other disasters, rescuers search for people buried under rubble by listening for the sounds which they make. Thus developing a technique to localize sound sources amidst loud noise will support such search and rescue operations. In this paper, we discuss an experiment performed to test an array signal processing technique which searches for unperceivable sound in loud noise environments. Two speakers simultaneously played a noise of a generator and a voice decreased by 20 dB (= 1/100 of power) from the generator noise at an outdoor space where cicadas were making noise. The sound signal was received by a horizontally set linear microphone array 1.05 m in length and consisting of 15 microphones. The direction and the distance of the voice were computed and the sound of the voice was extracted and played back as an audible sound by array signal processing.

  15. Artificially intelligent recognition of Arabic speaker using voice print-based local features

    NASA Astrophysics Data System (ADS)

    Mahmood, Awais; Alsulaiman, Mansour; Muhammad, Ghulam; Akram, Sheeraz

    2016-11-01

    Local features for any pattern recognition system are based on the information extracted locally. In this paper, a local feature extraction technique was developed. This feature was extracted in the time-frequency plain by taking the moving average on the diagonal directions of the time-frequency plane. This feature captured the time-frequency events producing a unique pattern for each speaker that can be viewed as a voice print of the speaker. Hence, we referred to this technique as voice print-based local feature. The proposed feature was compared to other features including mel-frequency cepstral coefficient (MFCC) for speaker recognition using two different databases. One of the databases used in the comparison is a subset of an LDC database that consisted of two short sentences uttered by 182 speakers. The proposed feature attained 98.35% recognition rate compared to 96.7% for MFCC using the LDC subset.

  16. A disturbance observer-based adaptive control approach for flexure beam nano manipulators.

    PubMed

    Zhang, Yangming; Yan, Peng; Zhang, Zhen

    2016-01-01

    This paper presents a systematic modeling and control methodology for a two-dimensional flexure beam-based servo stage supporting micro/nano manipulations. Compared with conventional mechatronic systems, such systems have major control challenges including cross-axis coupling, dynamical uncertainties, as well as input saturations, which may have adverse effects on system performance unless effectively eliminated. A novel disturbance observer-based adaptive backstepping-like control approach is developed for high precision servo manipulation purposes, which effectively accommodates model uncertainties and coupling dynamics. An auxiliary system is also introduced, on top of the proposed control scheme, to compensate the input saturations. The proposed control architecture is deployed on a customized-designed nano manipulating system featured with a flexure beam structure and voice coil actuators (VCA). Real time experiments on various manipulating tasks, such as trajectory/contour tracking, demonstrate precision errors of less than 1%. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.

  17. Human-computer interaction for alert warning and attention allocation systems of the multimodal watchstation

    NASA Astrophysics Data System (ADS)

    Obermayer, Richard W.; Nugent, William A.

    2000-11-01

    The SPAWAR Systems Center San Diego is currently developing an advanced Multi-Modal Watchstation (MMWS); design concepts and software from this effort are intended for transition to future United States Navy surface combatants. The MMWS features multiple flat panel displays and several modes of user interaction, including voice input and output, natural language recognition, 3D audio, stylus and gestural inputs. In 1999, an extensive literature review was conducted on basic and applied research concerned with alerting and warning systems. After summarizing that literature, a human computer interaction (HCI) designer's guide was prepared to support the design of an attention allocation subsystem (AAS) for the MMWS. The resultant HCI guidelines are being applied in the design of a fully interactive AAS prototype. An overview of key findings from the literature review, a proposed design methodology with illustrative examples, and an assessment of progress made in implementing the HCI designers guide are presented.

  18. A model for incorporating patient and stakeholder voices in a learning health care network: Washington State's Comparative Effectiveness Research Translation Network.

    PubMed

    Devine, Emily Beth; Alfonso-Cristancho, Rafael; Devlin, Allison; Edwards, Todd C; Farrokhi, Ellen T; Kessler, Larry; Lavallee, Danielle C; Patrick, Donald L; Sullivan, Sean D; Tarczy-Hornoch, Peter; Yanez, N David; Flum, David R

    2013-08-01

    To describe the inaugural comparative effectiveness research (CER) cohort study of Washington State's Comparative Effectiveness Research Translation Network (CERTAIN), which compares invasive with noninvasive treatments for peripheral artery disease, and to focus on the patient centeredness of this cohort study by describing it within the context of a newly published conceptual framework for patient-centered outcomes research (PCOR). The peripheral artery disease study was selected because of clinician-identified uncertainty in treatment selection and differences in desired outcomes between patients and clinicians. Patient centeredness is achieved through the "Patient Voices Project," a CERTAIN initiative through which patient-reported outcome (PRO) instruments are administered for research and clinical purposes, and a study-specific patient advisory group where patients are meaningfully engaged throughout the life cycle of the study. A clinician-led research advisory panel follows in parallel. Primary outcomes are PRO instruments that measure function, health-related quality of life, and symptoms, the latter developed with input from the patients. Input from the patient advisory group led to revised retention procedures, which now focus on short-term (3-6 months) follow-up. The research advisory panel is piloting a point-of-care, patient assessment checklist, thereby returning study results to practice. The cohort study is aligned with the tenets of one of the new conceptual frameworks for conducting PCOR. The CERTAIN's inaugural cohort study may serve as a useful model for conducting PCOR and creating a learning health care network. Copyright © 2013 Elsevier Inc. All rights reserved.

  19. A Model for Incorporating Patient and Stakeholder Voices in a Learning Healthcare Network: Washington State’s Comparative Effectiveness Research Translation Network (CERTAIN)

    PubMed Central

    Devine, EB; Alfonso-Cristancho, R; Devlin, A; Edwards, TC; Farrokhi, ET; Kessler, L; Lavallee, DC; Patrick, DL; Sullivan, SD; Tarczy-Hornoch, P; Yanez, ND; Flum, DR

    2014-01-01

    Objective To describe the inaugural comparative effectiveness research (CER) cohort study of Washington State’s Comparative Effectiveness Research Translation Network (CERTAIN), which compares invasive to non-invasive treatments for peripheral artery disease; to focus on the patient-centeredness of this cohort study by describing it within the context of a newly published conceptual frameworks for patient-centered outcomes research (PCOR). Study Design and Setting The peripheral artery disease study was selected due to clinician-identified uncertainty in treatment selection and differences in desired outcomes between patients and clinicians. Patient-centeredness is achieved through the ‘Patient Voices Project’, a CERTAIN initiative through which patient-reported outcome (PRO) instruments are administered for research and clinical purposes, and a study-specific patient advisory group where patients are meaningfully engaged throughout the life cycle of the trial. A clinician-led research advisory panel follows in parallel. Results Primary outcomes are PRO instruments that measure function, health-related quality of life, and symptoms; the latter developed with input from patients. Input from the patient advisory group led to revised retention procedures, which now focus on short-term (3–6 months) follow-up. The research advisory panel is piloting a point-of-care, patient assessment checklist, there by returning study results to practice. The cohort study is aligned with the tenets of one of the new conceptual frameworks for conducting PCOR. Conclusion CERTAIN’s inaugural cohort study may serve as a useful model for conducting PCOR and creating a Learning Healthcare Network. PMID:23849146

  20. Education Reform in New Orleans: Voices from the Recovery School District

    ERIC Educational Resources Information Center

    Ciolino, Max S.; Kirylo, James D.; Mirón, Luis; Frazier, Kelly

    2014-01-01

    In the post-Katrina education landscape in New Orleans, teachers in charter schools and district-run schools in the Recovery School District are uniquely situated to provide a direct eyewitness account of the successes and failures of the city's new direction in public education. This narrative presents the opinions of teachers in a critical…

  1. Maternal Sensitivity and the Learning-Promoting Effects of Depressed and Nondepressed Mothers' Infant-Directed Speech

    ERIC Educational Resources Information Center

    Kaplan, Peter S.; Burgess, Aaron P.; Sliter, Jessica K.; Moreno, Amanda J.

    2009-01-01

    The hypothesis that aspects of current mother-infant interactions predict an infant's response to maternal infant-directed speech (IDS) was tested. Relative to infants of nondepressed mothers, those of depressed mothers acquired weaker voice-face associations in response to their own mothers' IDS in a conditioned-attention paradigm, although this…

  2. Voices from the "Working Lives" Project: The Push-Pull of Work and Care

    ERIC Educational Resources Information Center

    Fehring, Heather; Herring, Katherine

    2012-01-01

    A recent policy direction in many OECD countries has been to increase workforce participation for women of childbearing age; a policy direction which seemingly runs counter to a need for improved work-life balance for women themselves. This article explores the impact of this somewhat contradictory "push-pull" of policy by examining some…

  3. Estimating the Intended Sound Direction of the User: Toward an Auditory Brain-Computer Interface Using Out-of-Head Sound Localization

    PubMed Central

    Nambu, Isao; Ebisawa, Masashi; Kogure, Masumi; Yano, Shohei; Hokari, Haruhide; Wada, Yasuhiro

    2013-01-01

    The auditory Brain-Computer Interface (BCI) using electroencephalograms (EEG) is a subject of intensive study. As a cue, auditory BCIs can deal with many of the characteristics of stimuli such as tone, pitch, and voices. Spatial information on auditory stimuli also provides useful information for a BCI. However, in a portable system, virtual auditory stimuli have to be presented spatially through earphones or headphones, instead of loudspeakers. We investigated the possibility of an auditory BCI using the out-of-head sound localization technique, which enables us to present virtual auditory stimuli to users from any direction, through earphones. The feasibility of a BCI using this technique was evaluated in an EEG oddball experiment and offline analysis. A virtual auditory stimulus was presented to the subject from one of six directions. Using a support vector machine, we were able to classify whether the subject attended the direction of a presented stimulus from EEG signals. The mean accuracy across subjects was 70.0% in the single-trial classification. When we used trial-averaged EEG signals as inputs to the classifier, the mean accuracy across seven subjects reached 89.5% (for 10-trial averaging). Further analysis showed that the P300 event-related potential responses from 200 to 500 ms in central and posterior regions of the brain contributed to the classification. In comparison with the results obtained from a loudspeaker experiment, we confirmed that stimulus presentation by out-of-head sound localization achieved similar event-related potential responses and classification performances. These results suggest that out-of-head sound localization enables us to provide a high-performance and loudspeaker-less portable BCI system. PMID:23437338

  4. 29 CFR 1926.1421 - Signals-voice signals-additional requirements.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... operations, the operator, signal person and lift director (if there is one), must contact each other and....), direction; distance and/or speed; function, stop command. (c) The operator, signal person and lift director...

  5. People-selectivity, audiovisual integration and heteromodality in the superior temporal sulcus.

    PubMed

    Watson, Rebecca; Latinus, Marianne; Charest, Ian; Crabbe, Frances; Belin, Pascal

    2014-01-01

    The functional role of the superior temporal sulcus (STS) has been implicated in a number of studies, including those investigating face perception, voice perception, and face-voice integration. However, the nature of the STS preference for these 'social stimuli' remains unclear, as does the location within the STS for specific types of information processing. The aim of this study was to directly examine properties of the STS in terms of selective response to social stimuli. We used functional magnetic resonance imaging (fMRI) to scan participants whilst they were presented with auditory, visual, or audiovisual stimuli of people or objects, with the intention of localising areas preferring both faces and voices (i.e., 'people-selective' regions) and audiovisual regions designed to specifically integrate person-related information. Results highlighted a 'people-selective, heteromodal' region in the trunk of the right STS which was activated by both faces and voices, and a restricted portion of the right posterior STS (pSTS) with an integrative preference for information from people, as compared to objects. These results point towards the dedicated role of the STS as a 'social-information processing' centre. Copyright © 2013 Elsevier Ltd. All rights reserved.

  6. Contemporary management of voice and swallowing disorders in patients with advanced lung cancer.

    PubMed

    Brady, Grainne C; Carding, Paul N; Bhosle, Jaishree; Roe, Justin W G

    2015-06-01

    Advanced lung cancer can cause changes to swallowing and communication function. Direct tumour invasion, dyspnoea and deconditioning can all impact on swallowing function and communication. Cancer treatment, if administered, may cause or compound symptoms. In this study, the nature of swallowing and communication difficulties in patients with advanced lung cancer will be discussed, and management options including medical management, speech and language therapy (SLT) intervention, and surgical interventions will be considered. Advanced lung cancer can result in voice and swallowing difficulties, which can increase symptom burden and significantly impact on quality of life (QOL). There is a growing evidence base to support the use of injection laryngoplasty under local anaesthetic to offer immediate improvement in voice, swallowing and overall QOL. There is limited literature on the nature and extent of voice and swallowing impairment in patients with lung cancer. Well designed studies with robust and sensitive multidimensional dysphagia and dysphonia assessments are required. Outcome studies examining interventions with clearly defined treatment goals are required. These studies should include both functional and patient-reported outcome measures to develop the evidence base and to ensure that interventions are both timely and appropriate.

  7. People-selectivity, audiovisual integration and heteromodality in the superior temporal sulcus

    PubMed Central

    Watson, Rebecca; Latinus, Marianne; Charest, Ian; Crabbe, Frances; Belin, Pascal

    2014-01-01

    The functional role of the superior temporal sulcus (STS) has been implicated in a number of studies, including those investigating face perception, voice perception, and face–voice integration. However, the nature of the STS preference for these ‘social stimuli’ remains unclear, as does the location within the STS for specific types of information processing. The aim of this study was to directly examine properties of the STS in terms of selective response to social stimuli. We used functional magnetic resonance imaging (fMRI) to scan participants whilst they were presented with auditory, visual, or audiovisual stimuli of people or objects, with the intention of localising areas preferring both faces and voices (i.e., ‘people-selective’ regions) and audiovisual regions designed to specifically integrate person-related information. Results highlighted a ‘people-selective, heteromodal’ region in the trunk of the right STS which was activated by both faces and voices, and a restricted portion of the right posterior STS (pSTS) with an integrative preference for information from people, as compared to objects. These results point towards the dedicated role of the STS as a ‘social-information processing’ centre. PMID:23988132

  8. The Accuracy of Preoperative Rigid Stroboscopy in the Evaluation of Voice Disorders in Children.

    PubMed

    Mansour, Jobran; Amir, Ofer; Sagiv, Doron; Alon, Eran E; Wolf, Michael; Primov-Fever, Adi

    2017-07-01

    Stroboscopy is considered the most appropriate tool for evaluating the function of the vocal folds but may harbor significant limitations in children. Still, direct laryngoscopy (DL), under general anesthesia, is regarded the "gold standard" for establishing a diagnosis of vocal fold pathology. The aim of the study is to examine the accuracy of preoperative rigid stroboscopy in children with voice disorders. This is a retrospective study. A retrospective study was conducted on a cohort of 39 children with dysphonia, aged 4 to 18 years, who underwent DL. Twenty-six children underwent rigid stroboscopy (RS) prior to surgery and 13 children underwent fiber-optic laryngoscopy. The preoperative diagnoses were matched with intraoperative (DL) findings. DL was found to contradict preoperative evaluations in 20 out of 39 children (51%) and in 26 out of 53 of the findings (49%). Overdiagnosis of cysts and underdiagnosis of sulci were noted in RS compared to DL. The overall rate of accuracy for RS was 64%. The accuracy of rigid stroboscopy in the evaluation of children with voice disorders was found to be similar with previous reports in adults. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  9. Nurses using futuristic technology in today's healthcare setting.

    PubMed

    Wolf, Debra M; Kapadia, Amar; Kintzel, Jessie; Anton, Bonnie B

    2009-01-01

    Human computer interaction (HCI) equates nurses using voice assisted technology within a clinical setting to document patient care real time, retrieve patient information from care plans, and complete routine tasks. This is a reality currently utilized by clinicians today in acute and long term care settings. Voice assisted documentation provides hands & eyes free accurate documentation while enabling effective communication and task management. The speech technology increases the accuracy of documentation, while interfacing directly into the electronic health record (EHR). Using technology consisting of a light weight headset and small fist size wireless computer, verbal responses to easy to follow cues are converted into a database systems allowing staff to obtain individualized care status reports on demand. To further assist staff in their daily process, this innovative technology allows staff to send and receive pages as needed. This paper will discuss how leading edge and award winning technology is being integrated within the United States. Collaborative efforts between clinicians and analyst will be discussed reflecting the interactive design and build functionality. Features such as the system's voice responses and directed cues will be shared and how easily data can be documented, viewed and retrieved. Outcome data will be presented on how the technology impacted organization's quality outcomes, financial reimbursement, and employee's level of satisfaction.

  10. En Route Air Traffic Control Input Devices for the Next Generation

    NASA Technical Reports Server (NTRS)

    Mainini, Matthew J.

    2010-01-01

    The purpose of this study was to investigate the usefulness of different input device configurations when trial planning new routes for aircraft in an advanced simulation of the en route workstation. The task of trial planning is one of the futuristic tools that is performed by the graphical manipulation of an aircraft's trajectory to reroute the aircraft without voice communication. In this study with two input devices, the FAA's current trackball and a basic optical computer mouse were evaluated with "pick" button in a click-and-hold state and a click-and-release state while the participant dragged the trial plan line. The trial plan was used for three different conflict types: Aircraft Conflicts, Weather Conflicts, and Aircraft + Weather Conflicts. Speed and accuracy were the primary dependent variables. Results indicate that the mouse conditions were significantly faster than the trackball conditions overall with no significant loss of accuracy. Several performance ratings and preference ratings were analyzed from post-run and post-simulation questionnaires. The release conditions were significantly more useful and likable than the hold conditions. The results suggest that the mouse in the release button state was the fastest and most well liked device configuration for trial planning in the en route workstation. Keywords-input devices, en route, controller, workstation, mouse, trackball, NextGen

  11. Language input and acquisition in a Mayan village: how important is directed speech?

    PubMed

    Shneidman, Laura A; Goldin-Meadow, Susan

    2012-09-01

    Theories of language acquisition have highlighted the importance of adult speakers as active participants in children's language learning. However, in many communities children are reported to be directly engaged by their caregivers only rarely (Lieven, 1994). This observation raises the possibility that these children learn language from observing, rather than participating in, communicative exchanges. In this paper, we quantify naturally occurring language input in one community where directed interaction with children has been reported to be rare (Yucatec Mayan). We compare this input to the input heard by children growing up in large families in the United States, and we consider how directed and overheard input relate to Mayan children's later vocabulary. In Study 1, we demonstrate that 1-year-old Mayan children do indeed hear a smaller proportion of total input in directed speech than children from the US. In Study 2, we show that for Mayan (but not US) children, there are great increases in the proportion of directed input that children receive between 13 and 35 months. In Study 3, we explore the validity of using videotaped data in a Mayan village. In Study 4, we demonstrate that word types directed to Mayan children from adults at 24 months (but not word types overheard by children or word types directed from other children) predict later vocabulary. These findings suggest that adult talk directed to children is important for early word learning, even in communities where much of children's early language input comes from overheard speech. © 2012 Blackwell Publishing Ltd.

  12. Native sound category formation in simultaneous bilingual acquisition

    NASA Astrophysics Data System (ADS)

    Bosch, Laura

    2004-05-01

    The consequences of early bilingual exposure on the perceptual reorganization processes that occur by the end of the first year of life were analyzed in a series of experiments on the capacity to discriminate vowel and consonant contrasts, comparing monolingual and bilingual infants (Catalan/Spanish) at different age levels. For bilingual infants, the discrimination of target vowel contrasts, which reflect different amount of overlapping and acoustic distance between the two languages of exposure, suggested a U-shaped developmental pattern. A similar trend was observed in the bilingual infants discrimination of a fricative voicing contrast, present in only one of the languages in their environment. The temporary decline in sensitivity found at 8 months for vowel targets and at 12 months for the voicing contrast reveals the specific perceptual processes that bilingual infants develop in order to deal with their complex linguistic input. Data from adult bilingual subjects on a lexical decision task involving these contrasts add to this developmental picture and suggest the existence of a dominant language even in simultaneous bilingual acquisition. [Work supported by JSMF 10001079BMB.

  13. New portable voice guidance device for the manual wheelchair transfer: a pilot study in patients with hemiplegia.

    PubMed

    Yoshida, Taiki; Otaka, Yohei; Osu, Rieko; Kita, Kahori; Sakata, Sachiko; Kondo, Kunitsugu

    2017-05-01

    Older and/or cognitively impaired patients require verbal guidance to prevent accidents during wheelchair operation, thus increasing the burden on caregivers. This study aimed to develop a new portable voice guidance device for manual wheelchairs and examine its clinical usefulness. We developed a portable voice guidance device to monitor the statuses of wheelchair brakes and footrests and automatically provide voice guidance for operation. The device comprises a microcomputer, four magnets and magnetic sensors, speaker and battery. Device operation was assessed during the transfer from a wheelchair to bed six times per day over three days for a total of 90 transfers in five stroke patients (mean age: 79.6 years) who required verbal guidance to direct wheelchair operation. Device usability was also assessed using a questionnaire. The device performed perfectly during all attempted transfers (100%). To ensure safety, the assessor needed to add verbal guidance during 33 of 90 attempted transfers (36.6%). Overall, the device usability was favourable. However, some assessors were unsatisfied with the volume of the device voice, guidance timing and burden reduction. Our device could facilitate wheelchair operation and might potentially be used to reduce fall risk in stroke patients and the burden on caregivers. Implications for Rehabilitation The acquisition of transfer independence is an important step in the rehabilitation of patients with mobility issues. Many patients require supervision and guidance regarding the operation of brakes and footrests on manual wheelchairs. This newly developed voice guidance device for manual wheelchair transfers worked well in patients with hemiplegia and might be helpful to reduce the fall risks and the burden of care.

  14. Survey alerts hospital to needs of consumers.

    PubMed

    Schoenfeldt, R C; Seale, W B; Hale, A W

    1987-09-01

    Because of rapidly changing developments in the healthcare field, more emphasis is being placed on marketing of hospital services. A hospital's success will depend more and more on strategic planning based on timely and accurate information. In light of this, Lourdes Hospital, Paducah, KY, undertook a survey to evaluate its current performance and to determine a path for the future. The survey found, among other discoveries, that patients want more voice in determining their own treatment; they prefer outpatient treatment when possible, even if it is not covered by insurance; and stress management and health assessment clinics are the most popular extra services a hospital could offer. Physicians surveyed said they wanted more input into the evaluation of new services and equipment at the hospital. The survey also found that most patients either select a hospital in conjunction with their physician or have their physician choose the hospital. The findings led to some major changes at the hospital, including a restructuring of the planning process to get physicians more involved, a new marketing strategy to enhance communication with consumers, and increased outpatient services. The results have given direction to the hospital administration, helped shape advertising, and provided support for certificate-of-need requests.

  15. Common neural systems associated with the recognition of famous faces and names: An event-related fMRI study

    PubMed Central

    Nielson, Kristy A.; Seidenberg, Michael; Woodard, John L.; Durgerian, Sally; Zhang, Qi; Gross, William L.; Gander, Amelia; Guidotti, Leslie M.; Antuono, Piero; Rao, Stephen M.

    2010-01-01

    Person recognition can be accomplished through several modalities (face, name, voice). Lesion, neurophysiology and neuroimaging studies have been conducted in an attempt to determine the similarities and differences in the neural networks associated with person identity via different modality inputs. The current study used event-related functional-MRI in 17 healthy participants to directly compare activation in response to randomly presented famous and non-famous names and faces (25 stimuli in each of the four categories). Findings indicated distinct areas of activation that differed for faces and names in regions typically associated with pre-semantic perceptual processes. In contrast, overlapping brain regions were activated in areas associated with the retrieval of biographical knowledge and associated social affective features. Specifically, activation for famous faces was primarily right lateralized and famous names were left lateralized. However, for both stimuli, similar areas of bilateral activity were observed in the early phases of perceptual processing. Activation for fame, irrespective of stimulus modality, activated an extensive left hemisphere network, with bilateral activity observed in the hippocampi, posterior cingulate, and middle temporal gyri. Findings are discussed within the framework of recent proposals concerning the neural network of person identification. PMID:20167415

  16. Dependence of phonatory effort on hydration level.

    PubMed

    Verdolini, K; Titze, I R; Fennell, A

    1994-10-01

    In this study, a double-blind placebo-controlled approach was used to assess the relation between hydration level and phonatory effort. Twelve adult, untrained voice users with normal voices participated as subjects. Each subject received a 4-hour hydration treatment, a 4-hour dehydration treatment, and a 4-hour placebo (control) treatment. Following each treatment, phonatory effort was measured with a physiological measure, phonation threshold pressure (PTP), and with a psychological measure, direct magnitude estimation of perceived phonatory effort (DMEPPE). Summarizing the results across these measures, the findings indicated an inverse relation between phonatory effort and hydration level, but primarily for high-pitched phonation tasks. The findings for PTPs replicated those from an earlier study conducted without double-blind experimental manipulations (Verdolini-Marston, Titze, & Druker, 1990). Theoretical discussion focuses on the possible role of vocal fold tissue viscosity for hydration and dehydration effects, although direct measures of tissue viscosity are lacking.

  17. Pathways to Youth Empowerment and Community Connectedness: A Study of Youth-Adult Partnership in Malaysian After-School, Co-Curricular Programs.

    PubMed

    Zeldin, Shepherd; Krauss, Steven Eric; Kim, Taehan; Collura, Jessica; Abdullah, Haslinda

    2016-08-01

    After-school programs are prevalent across the world, but there is a paucity of research that examines quality within the "black box" of programs at the point of service. Grounded in current theory, this research examined hypothesized pathways between the experience of youth-adult partnership (youth voice in decision-making; supportive adult relationships), the mediators of program safety and engagement, and the developmental outcomes of youth empowerment (leadership competence, policy control) and community connectedness (community connections, school attachment). Surveys were administered to 207 ethnically diverse (47.3 % female; 63.3 % Malay) youth, age 15-16, attending after-school co-curricular programs in Kuala Lumpur, Malaysia. Results showed that youth voice in program decision-making predicted both indicators of youth empowerment. Neither youth voice nor supportive adult relationships was directly associated with community connectedness, however. Program engagement mediated the associations between youth-adult partnership and empowerment. In contrast, program safety mediated the associations between youth-adult partnership and community connectedness. The findings indicate that the two core components of youth-adult partnership-youth voice and supportive adult relationships-may operate through different, yet complementary, pathways of program quality to predict developmental outcomes. Implications for future research are highlighted. For reasons of youth development and youth rights, the immediate challenge is to create opportunities for youth to speak on issues of program concern and to elevate those adults who are able and willing to help youth exercise their voice.

  18. Octopus Cells in the Posteroventral Cochlear Nucleus Provide the Main Excitatory Input to the Superior Paraolivary Nucleus

    PubMed Central

    Felix II, Richard A.; Gourévitch, Boris; Gómez-Álvarez, Marcelo; Leijon, Sara C. M.; Saldaña, Enrique; Magnusson, Anna K.

    2017-01-01

    Auditory streaming enables perception and interpretation of complex acoustic environments that contain competing sound sources. At early stages of central processing, sounds are segregated into separate streams representing attributes that later merge into acoustic objects. Streaming of temporal cues is critical for perceiving vocal communication, such as human speech, but our understanding of circuits that underlie this process is lacking, particularly at subcortical levels. The superior paraolivary nucleus (SPON), a prominent group of inhibitory neurons in the mammalian brainstem, has been implicated in processing temporal information needed for the segmentation of ongoing complex sounds into discrete events. The SPON requires temporally precise and robust excitatory input(s) to convey information about the steep rise in sound amplitude that marks the onset of voiced sound elements. Unfortunately, the sources of excitation to the SPON and the impact of these inputs on the behavior of SPON neurons have yet to be resolved. Using anatomical tract tracing and immunohistochemistry, we identified octopus cells in the contralateral cochlear nucleus (CN) as the primary source of excitatory input to the SPON. Cluster analysis of miniature excitatory events also indicated that the majority of SPON neurons receive one type of excitatory input. Precise octopus cell-driven onset spiking coupled with transient offset spiking make SPON responses well-suited to signal transitions in sound energy contained in vocalizations. Targets of octopus cell projections, including the SPON, are strongly implicated in the processing of temporal sound features, which suggests a common pathway that conveys information critical for perception of complex natural sounds. PMID:28620283

  19. Direct and octave-shifted pitch matching during nonword imitations in men, women, and children

    PubMed Central

    Peter, Beate; Foster, Bronsyn; Haas, Heather; Middleton, Kyle; McKibben, Kiersten

    2014-01-01

    Summary Objectives To evaluate whether children, women, and men match the speaker’s fundamental frequency (F0) during nonword imitation directly when the target F0 is within the responders’ vocal ranges and at octave-shifted levels when the target is outside their vocal ranges. To evaluate the role of a history of speech sound disorder (SSD) in the adult participants. Study Design Observational. Methods Nonword sets spoken by a man and a woman were imitated by 14 men, 21 women, and 19 children. Approximately half of the adults and two thirds of the children had a history of SSD. F0 in the imitations was compared to that in the targets and in the participants’ non-imitated control word productions. Results When the target F0 was within the responders’ vocal ranges, the imitations approximated the target F0. Men imitating a woman’s voice approximated F0 levels one octave below the target F0. Children imitating a man’s voice approximated F0 levels one octave above the target F0. Women imitating a man’s voice approximated the target F0 at a ratio of 1.5, known as the perfect fifth in music. A history of SSD did not influence the results. Conclusions This study replicates previous findings showing that target F0 was a salient aspect of the stimuli that was imitated along with the targets’ segmental and prosodic components without explicit prompting. It is the first to show F0 convergence not only directly but also at relevant target/imitation intervals including the octave interval. PMID:25439509

  20. Researchers fear 'Putin's Academy of Sciences'

    NASA Astrophysics Data System (ADS)

    Moskvitch, Katia

    2013-11-01

    Scientists have voiced concerns about the future of the Russian Academy of Sciences (RAS) after the country's president, Vladimir Putin, signed a law that will make the 289-year-old body come under the direct control of a new government agency.

  1. The Effects of Audiovisual Inputs on Solving the Cocktail Party Problem in the Human Brain: An fMRI Study.

    PubMed

    Li, Yuanqing; Wang, Fangyi; Chen, Yongbin; Cichocki, Andrzej; Sejnowski, Terrence

    2017-09-25

    At cocktail parties, our brains often simultaneously receive visual and auditory information. Although the cocktail party problem has been widely investigated under auditory-only settings, the effects of audiovisual inputs have not. This study explored the effects of audiovisual inputs in a simulated cocktail party. In our fMRI experiment, each congruent audiovisual stimulus was a synthesis of 2 facial movie clips, each of which could be classified into 1 of 2 emotion categories (crying and laughing). Visual-only (faces) and auditory-only stimuli (voices) were created by extracting the visual and auditory contents from the synthesized audiovisual stimuli. Subjects were instructed to selectively attend to 1 of the 2 objects contained in each stimulus and to judge its emotion category in the visual-only, auditory-only, and audiovisual conditions. The neural representations of the emotion features were assessed by calculating decoding accuracy and brain pattern-related reproducibility index based on the fMRI data. We compared the audiovisual condition with the visual-only and auditory-only conditions and found that audiovisual inputs enhanced the neural representations of emotion features of the attended objects instead of the unattended objects. This enhancement might partially explain the benefits of audiovisual inputs for the brain to solve the cocktail party problem. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  2. Changes in Voice Onset Time and Motor Speech Skills in Children following Motor Speech Therapy: Evidence from /pa/ productions

    PubMed Central

    Yu, Vickie Y.; Kadis, Darren S.; Oh, Anna; Goshulak, Debra; Namasivayam, Aravind; Pukonen, Margit; Kroll, Robert; De Nil, Luc F.; Pang, Elizabeth W.

    2016-01-01

    This study evaluated changes in motor speech control and inter-gestural coordination for children with speech sound disorders (SSD) subsequent to PROMPT (Prompts for Restructuring Oral Muscular Phonetic Targets) intervention. We measured the distribution patterns of voice onset time (VOT) for a voiceless stop (/p/) to examine the changes in inter-gestural coordination. Two standardized tests were used (VMPAC, GFTA-2) to assess the changes in motor speech skills and articulation. Data showed positive changes in patterns of VOT with a lower pattern of variability. All children showed significantly higher scores for VMPAC, but only some children showed higher scores for GFTA-2. Results suggest that the proprioceptive feedback provided through PROMPT had a positive influence on motor speech control and inter-gestural coordination in voicing behavior. This set of VOT data for children with SSD adds to our understanding of the speech characteristics underlying motor speech control. Directions for future studies are discussed. PMID:24446799

  3. A multiscale product approach for an automatic classification of voice disorders from endoscopic high-speed videos.

    PubMed

    Unger, Jakob; Schuster, Maria; Hecker, Dietmar J; Schick, Bernhard; Lohscheller, Joerg

    2013-01-01

    Direct observation of vocal fold vibration is indispensable for a clinical diagnosis of voice disorders. Among current imaging techniques, high-speed videoendoscopy constitutes a state-of-the-art method capturing several thousand frames per second of the vocal folds during phonation. Recently, a method for extracting descriptive features from phonovibrograms, a two-dimensional image containing the spatio-temporal pattern of vocal fold dynamics, was presented. The derived features are closely related to a clinically established protocol for functional assessment of pathologic voices. The discriminative power of these features for different pathologic findings and configurations has not been assessed yet. In the current study, a collective of 220 subjects is considered for two- and multi-class problems of healthy and pathologic findings. The performance of the proposed feature set is compared to conventional feature reduction routines and was found to clearly outperform these. As such, the proposed procedure shows great potential for diagnostical issues of vocal fold disorders.

  4. The Sound of Intellect: Speech Reveals a Thoughtful Mind, Increasing a Job Candidate's Appeal.

    PubMed

    Schroeder, Juliana; Epley, Nicholas

    2015-06-01

    A person's mental capacities, such as intellect, cannot be observed directly and so are instead inferred from indirect cues. We predicted that a person's intellect would be conveyed most strongly through a cue closely tied to actual thinking: his or her voice. Hypothetical employers (Experiments 1-3b) and professional recruiters (Experiment 4) watched, listened to, or read job candidates' pitches about why they should be hired. These evaluators rated a candidate as more competent, thoughtful, and intelligent when they heard a pitch rather than read it and, as a result, had a more favorable impression of the candidate and were more interested in hiring the candidate. Adding voice to written pitches, by having trained actors (Experiment 3a) or untrained adults (Experiment 3b) read them, produced the same results. Adding visual cues to audio pitches did not alter evaluations of the candidates. For conveying one's intellect, it is important that one's voice, quite literally, be heard. © The Author(s) 2015.

  5. The effects of a song-singing programme on the affective speaking intonation of people with traumatic brain injury.

    PubMed

    Baker, F; Wigram, T; Gold, C

    2005-07-01

    To examine changes in the relationship between intonation, voice range and mood following music therapy programmes in people with traumatic brain injury. Data from four case studies were pooled and effect size, ANOVA and correlation calculations were performed to evaluate the effectiveness of treatment. Subjects sang three self-selected songs for 15 sessions. Speaking fundamental frequency, fundamental frequency variability, slope, voice range and mood were analysed pre- and post-session. Immediate treatment effects were not found. Long-term improvements in affective intonation were found in three subjects, especially in fundamental frequency. Voice range improved over time and was positively correlated with the three intonation components. Mood scale data showed that immediate effects were in the negative direction whereas there weres increases in positive mood state in the longer-term. Findings suggest that, in the long-term, song singing can improve vocal range and mood and enhance the affective intonation styles of people with TBI.

  6. The Effectiveness of Pitch-raising Surgery in Male-to-Female Transsexuals: A Systematic Review.

    PubMed

    Van Damme, Silke; Cosyns, Marjan; Deman, Sofie; Van den Eede, Zoë; Van Borsel, John

    2017-03-01

    This study aimed to review the evidence of the effectiveness of pitch-raising surgery performed in male-to-female transsexuals. A search for studies was performed in PubMed, Web of Science, Science Direct, EBSCOhost, Google Scholar, and the references in retrieved manuscripts, using as keywords "transsexual" or "transgender" combined with terms related to voice surgery. We included eight studies using cricothyroid approximation, six studies using anterior glottal web formation, and six studies using other surgery types or a combination of surgical techniques, leading to 20 studies in total. Objectively, a substantial rise in postoperative fundamental frequency was identified. Perceptually, mainly laryngeal web formation seems risky for decreasing voice quality. The majority of patients seemed satisfied with the outcome. However, none of the studies used a control group and randomization process. Further investigation regarding long-term results is necessary. Future research needs to investigate long-term effects of pitch-raising surgery using a stronger study design. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  7. Rapid change in articulatory lip movement induced by preceding auditory feedback during production of bilabial plosives.

    PubMed

    Mochida, Takemi; Gomi, Hiroaki; Kashino, Makio

    2010-11-08

    There has been plentiful evidence of kinesthetically induced rapid compensation for unanticipated perturbation in speech articulatory movements. However, the role of auditory information in stabilizing articulation has been little studied except for the control of voice fundamental frequency, voice amplitude and vowel formant frequencies. Although the influence of auditory information on the articulatory control process is evident in unintended speech errors caused by delayed auditory feedback, the direct and immediate effect of auditory alteration on the movements of articulators has not been clarified. This work examined whether temporal changes in the auditory feedback of bilabial plosives immediately affects the subsequent lip movement. We conducted experiments with an auditory feedback alteration system that enabled us to replace or block speech sounds in real time. Participants were asked to produce the syllable /pa/ repeatedly at a constant rate. During the repetition, normal auditory feedback was interrupted, and one of three pre-recorded syllables /pa/, /Φa/, or /pi/, spoken by the same participant, was presented once at a different timing from the anticipated production onset, while no feedback was presented for subsequent repetitions. Comparisons of the labial distance trajectories under altered and normal feedback conditions indicated that the movement quickened during the short period immediately after the alteration onset, when /pa/ was presented 50 ms before the expected timing. Such change was not significant under other feedback conditions we tested. The earlier articulation rapidly induced by the progressive auditory input suggests that a compensatory mechanism helps to maintain a constant speech rate by detecting errors between the internally predicted and actually provided auditory information associated with self movement. The timing- and context-dependent effects of feedback alteration suggest that the sensory error detection works in a temporally asymmetric window where acoustic features of the syllable to be produced may be coded.

  8. A Case Series of the Probability Density and Cumulative Distribution of Laryngeal Disease in a Tertiary Care Voice Center.

    PubMed

    de la Fuente, Jaime; Garrett, C Gaelyn; Ossoff, Robert; Vinson, Kim; Francis, David O; Gelbard, Alexander

    2017-11-01

    To examine the distribution of clinic and operative pathology in a tertiary care laryngology practice. Probability density and cumulative distribution analyses (Pareto analysis) was used to rank order laryngeal conditions seen in an outpatient tertiary care laryngology practice and those requiring surgical intervention during a 3-year period. Among 3783 new clinic consultations and 1380 operative procedures, voice disorders were the most common primary diagnostic category seen in clinic (n = 3223), followed by airway (n = 374) and swallowing (n = 186) disorders. Within the voice strata, the most common primary ICD-9 code used was dysphonia (41%), followed by unilateral vocal fold paralysis (UVFP) (9%) and cough (7%). Among new voice patients, 45% were found to have a structural abnormality. The most common surgical indications were laryngotracheal stenosis (37%), followed by recurrent respiratory papillomatosis (18%) and UVFP (17%). Nearly 55% of patients presenting to a tertiary referral laryngology practice did not have an identifiable structural abnormality in the larynx on direct or indirect examination. The distribution of ICD-9 codes requiring surgical intervention was disparate from that seen in clinic. Application of the Pareto principle may improve resource allocation in laryngology, but these initial results require confirmation across multiple institutions.

  9. Integration of Motor Learning Principles Into Real-Time Ambulatory Voice Biofeedback and Example Implementation Via a Clinical Case Study With Vocal Fold Nodules.

    PubMed

    Van Stan, Jarrad H; Mehta, Daryush D; Petit, Robert J; Sternad, Dagmar; Muise, Jason; Burns, James A; Hillman, Robert E

    2017-02-01

    Ambulatory voice biofeedback (AVB) has the potential to significantly improve voice therapy effectiveness by targeting one of the most challenging aspects of rehabilitation: carryover of desired behaviors outside of the therapy session. Although initial evidence indicates that AVB can alter vocal behavior in daily life, retention of the new behavior after biofeedback has not been demonstrated. Motor learning studies repeatedly have shown retention-related benefits when reducing feedback frequency or providing summary statistics. Therefore, novel AVB settings that are based on these concepts are developed and implemented. The underlying theoretical framework and resultant implementation of innovative AVB settings on a smartphone-based voice monitor are described. A clinical case study demonstrates the functionality of the new relative frequency feedback capabilities. With new technical capabilities, 2 aspects of feedback are directly modifiable for AVB: relative frequency and summary feedback. Although reduced-frequency AVB was associated with improved carryover of a therapeutic vocal behavior (i.e., reduced vocal intensity) in a patient post-excision of vocal fold nodules, causation cannot be assumed. Timing and frequency of AVB schedules can be manipulated to empirically assess generalization of motor learning principles to vocal behavior modification and test the clinical effectiveness of AVB with various feedback schedules.

  10. Integration of Motor Learning Principles Into Real-Time Ambulatory Voice Biofeedback and Example Implementation Via a Clinical Case Study With Vocal Fold Nodules

    PubMed Central

    Mehta, Daryush D.; Petit, Robert J.; Sternad, Dagmar; Muise, Jason; Burns, James A.; Hillman, Robert E.

    2017-01-01

    Purpose Ambulatory voice biofeedback (AVB) has the potential to significantly improve voice therapy effectiveness by targeting one of the most challenging aspects of rehabilitation: carryover of desired behaviors outside of the therapy session. Although initial evidence indicates that AVB can alter vocal behavior in daily life, retention of the new behavior after biofeedback has not been demonstrated. Motor learning studies repeatedly have shown retention-related benefits when reducing feedback frequency or providing summary statistics. Therefore, novel AVB settings that are based on these concepts are developed and implemented. Method The underlying theoretical framework and resultant implementation of innovative AVB settings on a smartphone-based voice monitor are described. A clinical case study demonstrates the functionality of the new relative frequency feedback capabilities. Results With new technical capabilities, 2 aspects of feedback are directly modifiable for AVB: relative frequency and summary feedback. Although reduced-frequency AVB was associated with improved carryover of a therapeutic vocal behavior (i.e., reduced vocal intensity) in a patient post-excision of vocal fold nodules, causation cannot be assumed. Conclusions Timing and frequency of AVB schedules can be manipulated to empirically assess generalization of motor learning principles to vocal behavior modification and test the clinical effectiveness of AVB with various feedback schedules. PMID:28124070

  11. 47 CFR 27.1203 - EBS programming requirements.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... MISCELLANEOUS WIRELESS COMMUNICATIONS SERVICES Broadband Radio Service and Educational Broadband Service § 27... Broadband Service stations are intended primarily through video, data, or voice transmissions to further the... endeavors; (2) Transmission of material directly related to the administrative activities of the licensee...

  12. Knowledge Discovery, Integration and Communication for Extreme Weather and Flood Resilience Using Artificial Intelligence: Flood AI Alpha

    NASA Astrophysics Data System (ADS)

    Demir, I.; Sermet, M. Y.

    2016-12-01

    Nobody is immune from extreme events or natural hazards that can lead to large-scale consequences for the nation and public. One of the solutions to reduce the impacts of extreme events is to invest in improving resilience with the ability to better prepare, plan, recover, and adapt to disasters. The National Research Council (NRC) report discusses the topic of how to increase resilience to extreme events through a vision of resilient nation in the year 2030. The report highlights the importance of data, information, gaps and knowledge challenges that needs to be addressed, and suggests every individual to access the risk and vulnerability information to make their communities more resilient. This abstracts presents our project on developing a resilience framework for flooding to improve societal preparedness with objectives; (a) develop a generalized ontology for extreme events with primary focus on flooding; (b) develop a knowledge engine with voice recognition, artificial intelligence, natural language processing, and inference engine. The knowledge engine will utilize the flood ontology and concepts to connect user input to relevant knowledge discovery outputs on flooding; (c) develop a data acquisition and processing framework from existing environmental observations, forecast models, and social networks. The system will utilize the framework, capabilities and user base of the Iowa Flood Information System (IFIS) to populate and test the system; (d) develop a communication framework to support user interaction and delivery of information to users. The interaction and delivery channels will include voice and text input via web-based system (e.g. IFIS), agent-based bots (e.g. Microsoft Skype, Facebook Messenger), smartphone and augmented reality applications (e.g. smart assistant), and automated web workflows (e.g. IFTTT, CloudWork) to open the knowledge discovery for flooding to thousands of community extensible web workflows.

  13. Does insecure attachment mediate the relationship between trauma and voice-hearing in psychosis?

    PubMed

    Pilton, Marie; Bucci, Sandra; McManus, James; Hayward, Mark; Emsley, Richard; Berry, Katherine

    2016-12-30

    This study extends existing research and theoretical developments by exploring the potential mediating role of insecure attachment within the relationship between trauma and voice-hearing. Fifty-five voice hearers with a psychosis-related diagnosis completed comprehensive assessments of childhood trauma, adult attachment, voice-related severity and distress, beliefs about voices and relationships with voices. Anxious attachment was significantly associated with the voice-hearing dimensions examined. More sophisticated analysis showed that anxious attachment mediated the relationship between childhood sexual and emotional abuse and voice-related severity and distress, voice-malevolence, voice-omnipotence, voice-resistance and hearer-dependence. Anxious attachment also mediated the relationship between childhood physical neglect and voice-related severity and distress and hearer-dependence. Furthermore, consistent with previous research, the relationship between anxious attachment and voice-related distress was mediated by voice-malevolence, voice-omnipotence and voice-resistance. We propose a model whereby anxious attachment mediates the well-established relationship between trauma and voice-hearing. In turn, negative beliefs about voices may mediate the association between anxious attachment and voice-related distress. Findings presented here highlight the need to assess and formulate the impact of attachment patterns upon the voice-hearing experience in psychosis and the potential to alleviate voice-related distress by fostering secure attachments to therapists or significant others. Crown Copyright © 2016. Published by Elsevier Ireland Ltd. All rights reserved.

  14. How do teachers with self-reported voice problems differ from their peers with self-reported voice health?

    PubMed

    Lyberg Åhlander, Viveka; Rydell, Roland; Löfqvist, Anders

    2012-07-01

    This randomized case-control study compares teachers with self-reported voice problems to age-, gender-, and school-matched colleagues with self-reported voice health. The self-assessed voice function is related to factors known to influence the voice: laryngeal findings, voice quality, personality, psychosocial and coping aspects, searching for causative factors of voice problems in teachers. Subjects and controls, recruited from a teacher group in an earlier questionnaire study, underwent examinations of the larynx by high-speed imaging and kymograms; voice recordings; voice range profile; audiometry; self-assessment of voice handicap and voice function; teaching and environmental aspects; personality; coping; burnout, and work-related issues. The laryngeal and voice recordings were assessed by experienced phoniatricians and speech pathologists. The subjects with self-assessed voice problems differed from their peers with self-assessed voice health by significantly longer recovery time from voice problems and scored higher on all subscales of the Voice Handicap Index-Throat. The results show that the cause of voice dysfunction in this group of teachers with self-reported voice problems is not found in the vocal apparatus or within the individual. The individual's perception of a voice problem seems to be based on a combination of the number of symptoms and of how often the symptoms occur, along with the recovery time. The results also underline the importance of using self-assessed reports of voice dysfunction. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  15. Interpersonal Processes and Attachment in Voice-Hearers.

    PubMed

    Robson, George; Mason, Oliver

    2015-11-01

    Studies of both clinical and non-clinical voice hearers suggest that distress is rather inconsistently associated with the perceived relationship between voice and hearer. It is also not clear if their beliefs about voices are relevant. This study investigated the links between attachment anxiety/avoidance, interpersonal aspects of the voice relationship, and distress whilst considering the impact of beliefs about voices and paranoia. Forty-four voice-hearing participants completed a number of self-report measures tapping attachment, interpersonal processes in the voice relationship, beliefs about voices, paranoia, distress and depression. Attachment avoidance was related to voice intrusiveness, hearer distance and distress. Attachment anxiety was related to voice intrusiveness, hearer dependence and distress. A series of simple mediation analyses were conducted that suggest that the relationship between attachment and voice related distress may be mediated by interpersonal dynamics in the voice-hearer relationship, beliefs about voices and paranoia. Beliefs about voices, the hearer's relationship with their voices, and the distress voices sometimes engender appear to be meaningfully related to their attachment style. This may be important to consider in therapeutic work.

  16. Connections between voice ergonomic risk factors and voice symptoms, voice handicap, and respiratory tract diseases.

    PubMed

    Rantala, Leena M; Hakala, Suvi J; Holmqvist, Sofia; Sala, Eeva

    2012-11-01

    The aim of the study was to investigate the connections between voice ergonomic risk factors found in classrooms and voice-related problems in teachers. Voice ergonomic assessment was performed in 39 classrooms in 14 elementary schools by means of a Voice Ergonomic Assessment in Work Environment--Handbook and Checklist. The voice ergonomic risk factors assessed included working culture, noise, indoor air quality, working posture, stress, and access to a sound amplifier. Teachers from the above-mentioned classrooms reported their voice symptoms, respiratory tract diseases, and completed a Voice Handicap Index (VHI). The more voice ergonomic risk factors found in the classroom the higher were the teachers' total scores on voice symptoms and VHI. Stress was the factor that correlated most strongly with voice symptoms. Poor indoor air quality increased the occurrence of laryngitis. Voice ergonomics were poor in the classrooms studied and voice ergonomic risk factors affected the voice. It is important to convey information on voice ergonomics to education administrators and those responsible for school planning and taking care of school buildings. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  17. Comparison of a personalized parent voice smoke alarm with a conventional residential tone smoke alarm for awakening children.

    PubMed

    Smith, Gary A; Splaingard, Mark; Hayes, John R; Xiang, Huiyun

    2006-10-01

    Conventional residential tone smoke alarms fail to awaken the majority of children during slow wave sleep. With the objective of identifying a more effective smoke alarm for children, we compared a personalized parent voice smoke alarm with a conventional residential tone smoke alarm, both presented at 100 dB, with respect to their ability to awaken children 6- to 12-years-old from stage 4 sleep and prompt their performance of a simulated self-rescue escape procedure. Using a randomized, nonblinded, clinical research design, a volunteer sample of healthy children 6- to 12-years-old was enrolled in the study. Children were trained how to perform a simulated self-rescue escape procedure when they heard a smoke alarm. Each child's mother recorded a voice alarm message, "First name! First name! Wake up! Get out of bed! Leave the room!" For each child, either the voice or tone smoke alarm was randomly selected and triggered during the first cycle of stage 4 sleep, and then the other alarm was triggered during the second cycle of stage 4 sleep. Children's sleep stage was monitored by electroencephalography, electro-oculography, and chin electromyography. The 4 main outcome measures included the number of children who awakened, the number of children who escaped, the time to awakening, and the time to escape. Twenty-four children were enrolled. The median age was 9 years, and 11 (46%) were boys. One half of the children received the parent voice alarm first, and one half received the tone alarm first; however, the order that the alarm stimuli were presented was not statistically associated with awakening or escaping. Twenty-three (96%) of the 24 subjects awakened to the parent voice alarm compared with 14 (58%) to the tone alarm. One child did not awaken to either stimulus. Nine children awakened to their parent's voice but not to the tone, whereas none awakened to only the tone and not the voice. Twenty (83%) of the subjects in the parent voice alarm group successfully performed the escape procedure within 5 minutes of alarm onset compared with 9 (38%) in the tone alarm group. The median time to awaken was 20 seconds in the voice alarm group compared with 3 minutes in the tone alarm group. The median time to escape was 38 seconds in the voice alarm group compared with the maximum allowed 5 minutes in the tone alarm group. When exposed to the tone alarm, older children were more likely to awaken and were more likely to escape than younger children. There was no association between child's age and awakening or escaping for children exposed to the parent voice alarm. There was no association between child's gender and awakening or escaping for either alarm type. To our knowledge, this study is the first to compare the ability of different types of smoke alarms to awaken children while monitoring sleep stage. The personalized parent voice smoke alarm at 100 dB successfully awakened 96% of children 6- to 12-years-old from stage 4 sleep with 83% successfully performing a simulated self-rescue escape procedure, significantly outperforming the 100-dB conventional residential tone smoke alarm. These findings suggest a clear direction for future research, as well as important fundamental changes in smoke alarm design, that address the unique developmental needs of children. The development of a more effective smoke alarm for use in homes and other locations where children sleep provides an opportunity to reduce fire-related morbidity and mortality among children.

  18. Selective attention modulates early human evoked potentials during emotional face-voice processing.

    PubMed

    Ho, Hao Tam; Schröger, Erich; Kotz, Sonja A

    2015-04-01

    Recent findings on multisensory integration suggest that selective attention influences cross-sensory interactions from an early processing stage. Yet, in the field of emotional face-voice integration, the hypothesis prevails that facial and vocal emotional information interacts preattentively. Using ERPs, we investigated the influence of selective attention on the perception of congruent versus incongruent combinations of neutral and angry facial and vocal expressions. Attention was manipulated via four tasks that directed participants to (i) the facial expression, (ii) the vocal expression, (iii) the emotional congruence between the face and the voice, and (iv) the synchrony between lip movement and speech onset. Our results revealed early interactions between facial and vocal emotional expressions, manifested as modulations of the auditory N1 and P2 amplitude by incongruent emotional face-voice combinations. Although audiovisual emotional interactions within the N1 time window were affected by the attentional manipulations, interactions within the P2 modulation showed no such attentional influence. Thus, we propose that the N1 and P2 are functionally dissociated in terms of emotional face-voice processing and discuss evidence in support of the notion that the N1 is associated with cross-sensory prediction, whereas the P2 relates to the derivation of an emotional percept. Essentially, our findings put the integration of facial and vocal emotional expressions into a new perspective-one that regards the integration process as a composite of multiple, possibly independent subprocesses, some of which are susceptible to attentional modulation, whereas others may be influenced by additional factors.

  19. Native voice, self-concept and the moral case for personalized voice technology.

    PubMed

    Nathanson, Esther

    2017-01-01

    Purpose (1) To explore the role of native voice and effects of voice loss on self-concept and identity, and survey the state of assistive voice technology; (2) to establish the moral case for developing personalized voice technology. Methods This narrative review examines published literature on the human significance of voice, the impact of voice loss on self-concept and identity, and the strengths and limitations of current voice technology. Based on the impact of voice loss on self and identity, and voice technology limitations, the moral case for personalized voice technology is developed. Results Given the richness of information conveyed by voice, loss of voice constrains expression of the self, but the full impact is poorly understood. Augmentative and alternative communication (AAC) devices facilitate communication but, despite advances in this field, voice output cannot yet express the unique nuances of individual voice. The ethical principles of autonomy, beneficence and equality of opportunity establish the moral responsibility to invest in accessible, cost-effective, personalized voice technology. Conclusions Although further research is needed to elucidate the full effects of voice loss on self-concept, identity and social functioning, current understanding of the profoundly negative impact of voice loss establishes the moral case for developing personalized voice technology. Implications for Rehabilitation Rehabilitation of voice-disordered patients should facilitate self-expression, interpersonal connectedness and social/occupational participation. Proactive questioning about the psychological and social experiences of patients with voice loss is a valuable entry point for rehabilitation planning. Personalized voice technology would enhance sense of self, communicative participation and autonomy and promote shared healthcare decision-making. Further research is needed to identify the best strategies to preserve and strengthen identity and sense of self.

  20. Vocal Age Disguise: The Role of Fundamental Frequency and Speech Rate and Its Perceived Effects.

    PubMed

    Skoog Waller, Sara; Eriksson, Mårten

    2016-01-01

    The relationship between vocal characteristics and perceived age is of interest in various contexts, as is the possibility to affect age perception through vocal manipulation. A few examples of such situations are when age is staged by actors, when ear witnesses make age assessments based on vocal cues only or when offenders (e.g., online groomers) disguise their voice to appear younger or older. This paper investigates how speakers spontaneously manipulate two age related vocal characteristics ( f 0 and speech rate) in attempt to sound younger versus older than their true age, and if the manipulations correspond to actual age related changes in f 0 and speech rate (Study 1). Further aims of the paper is to determine how successful vocal age disguise is by asking listeners to estimate the age of generated speech samples (Study 2) and to examine whether or not listeners use f 0 and speech rate as cues to perceived age. In Study 1, participants from three age groups (20-25, 40-45, and 60-65 years) agreed to read a short text under three voice conditions. There were 12 speakers in each age group (six women and six men). They used their natural voice in one condition, attempted to sound 20 years younger in another and 20 years older in a third condition. In Study 2, 60 participants (listeners) listened to speech samples from the three voice conditions in Study 1 and estimated the speakers' age. Each listener was exposed to all three voice conditions. The results from Study 1 indicated that the speakers increased fundamental frequency ( f 0 ) and speech rate when attempting to sound younger and decreased f 0 and speech rate when attempting to sound older. Study 2 showed that the voice manipulations had an effect in the sought-after direction, although the achieved mean effect was only 3 years, which is far less than the intended effect of 20 years. Moreover, listeners used speech rate, but not f 0 , as a cue to speaker age. It was concluded that age disguise by voice can be achieved by naïve speakers even though the perceived effect was smaller than intended.

  1. Auditory traits of "own voice".

    PubMed

    Kimura, Marino; Yotsumoto, Yuko

    2018-01-01

    People perceive their recorded voice differently from their actively spoken voice. The uncanny valley theory proposes that as an object approaches humanlike characteristics, there is an increase in the sense of familiarity; however, eventually a point is reached where the object becomes strangely similar and makes us feel uneasy. The feeling of discomfort experienced when people hear their recorded voice may correspond to the floor of the proposed uncanny valley. To overcome the feeling of eeriness of own-voice recordings, previous studies have suggested equalization of the recorded voice with various types of filters, such as step, bandpass, and low-pass, yet the effectiveness of these filters has not been evaluated. To address this, the aim of experiment 1 was to identify what type of voice recording was the most representative of one's own voice. The voice recordings were presented in five different conditions: unadjusted recorded voice, step filtered voice, bandpass filtered voice, low-pass filtered voice, and a voice for which the participants freely adjusted the parameters. We found large individual differences in the most representative own-voice filter. In order to consider roles of sense of agency, experiment 2 investigated if lip-synching would influence the rating of own voice. The result suggested lip-synching did not affect own voice ratings. In experiment 3, based on the assumption that the voices used in previous experiments corresponded to continuous representations of non-own voice to own voice, the existence of an uncanny valley was examined. Familiarity, eeriness, and the sense of own voice were rated. The result did not support the existence of an uncanny valley. Taken together, the experiments led us to the following conclusions: there is no general filter that can represent own voice for everyone, sense of agency has no effect on own voice rating, and the uncanny valley does not exist for own voice, specifically.

  2. Forecast Modelling via Variations in Binary Image-Encoded Information Exploited by Deep Learning Neural Networks.

    PubMed

    Liu, Da; Xu, Ming; Niu, Dongxiao; Wang, Shoukai; Liang, Sai

    2016-01-01

    Traditional forecasting models fit a function approximation from dependent invariables to independent variables. However, they usually get into trouble when date are presented in various formats, such as text, voice and image. This study proposes a novel image-encoded forecasting method that input and output binary digital two-dimensional (2D) images are transformed from decimal data. Omitting any data analysis or cleansing steps for simplicity, all raw variables were selected and converted to binary digital images as the input of a deep learning model, convolutional neural network (CNN). Using shared weights, pooling and multiple-layer back-propagation techniques, the CNN was adopted to locate the nexus among variations in local binary digital images. Due to the computing capability that was originally developed for binary digital bitmap manipulation, this model has significant potential for forecasting with vast volume of data. The model was validated by a power loads predicting dataset from the Global Energy Forecasting Competition 2012.

  3. Listening to the student voice to improve educational software.

    PubMed

    van Wyk, Mari; van Ryneveld, Linda

    2017-01-01

    Academics often develop software for teaching and learning purposes with the best of intentions, only to be disappointed by the low acceptance rate of the software by their students once it is implemented. In this study, the focus is on software that was designed to enable veterinary students to record their clinical skills. A pilot of the software clearly showed that the program had not been received as well as had been anticipated, and therefore the researchers used a group interview and a questionnaire with closed-ended and open-ended questions to obtain the students' feedback. The open-ended questions were analysed with conceptual content analysis, and themes were identified. Students made valuable suggestions about what they regarded as important considerations when a new software program is introduced. The most important lesson learnt was that students cannot always predict their needs accurately if they are asked for input prior to the development of software. For that reason student input should be obtained on a continuous and regular basis throughout the design and development phases.

  4. Communication system with adaptive noise suppression

    NASA Technical Reports Server (NTRS)

    Kozel, David (Inventor); Devault, James A. (Inventor); Birr, Richard B. (Inventor)

    2007-01-01

    A signal-to-noise ratio dependent adaptive spectral subtraction process eliminates noise from noise-corrupted speech signals. The process first pre-emphasizes the frequency components of the input sound signal which contain the consonant information in human speech. Next, a signal-to-noise ratio is determined and a spectral subtraction proportion adjusted appropriately. After spectral subtraction, low amplitude signals can be squelched. A single microphone is used to obtain both the noise-corrupted speech and the average noise estimate. This is done by determining if the frame of data being sampled is a voiced or unvoiced frame. During unvoiced frames an estimate of the noise is obtained. A running average of the noise is used to approximate the expected value of the noise. Spectral subtraction may be performed on a composite noise-corrupted signal, or upon individual sub-bands of the noise-corrupted signal. Pre-averaging of the input signal's magnitude spectrum over multiple time frames may be performed to reduce musical noise.

  5. Forecast Modelling via Variations in Binary Image-Encoded Information Exploited by Deep Learning Neural Networks

    PubMed Central

    Xu, Ming; Niu, Dongxiao; Wang, Shoukai; Liang, Sai

    2016-01-01

    Traditional forecasting models fit a function approximation from dependent invariables to independent variables. However, they usually get into trouble when date are presented in various formats, such as text, voice and image. This study proposes a novel image-encoded forecasting method that input and output binary digital two-dimensional (2D) images are transformed from decimal data. Omitting any data analysis or cleansing steps for simplicity, all raw variables were selected and converted to binary digital images as the input of a deep learning model, convolutional neural network (CNN). Using shared weights, pooling and multiple-layer back-propagation techniques, the CNN was adopted to locate the nexus among variations in local binary digital images. Due to the computing capability that was originally developed for binary digital bitmap manipulation, this model has significant potential for forecasting with vast volume of data. The model was validated by a power loads predicting dataset from the Global Energy Forecasting Competition 2012. PMID:27281032

  6. Overall voice and strain level analysis in rock singers.

    PubMed

    Gonsalves, Aline; Amin, Elisabeth; Behlau, Mara

    2010-01-01

    overall voice and strain level analysis in rock singers. to analyze the voice o rock singers according to two specific parameters: overall level of vocal deviation (OLVD) and strain level (SL); to compare these parameters in three different music samples. participants were 26 male rock singers, ranging in age from 17 to 46 years (mean = 29.8 years). All of the participants answered a questionnaire for sample characterization and were submitted to the recording of three voice samples: Brazilian National Anthem (BNA), Satisfaction and self-selected repertoire song (RS). Voice samples were analyzed by five speech-language pathologists according to OLVD and SL. Statistical analysis was done using the software SPSS, version 13.0. statistically significant differences were observed for the mean values of OLVD and SL during the performance of Satisfaction (OLVD = 32.8 and SL = 0.024 / p=0.024) and during the RS performance (OLVD = 38.4 and SL = 55.8 / p=0.010). The values of OLVD and SL are directly proportional to the samples of the BNA* and RS**, i.e. the higher the strain the higher the OLVD (p,0.001*; p=0.010**). When individually analyzing the three song samples, it is observed that the OLVD does not vary significantly among them. However, the mean values present a trend to increase from non-rock to rock performances (24.0 BNA / 32.8 Satisfaction / 38.4 RS). The level of strain found during the BNA performance presents statistically significant difference when compared to the rock performances (Satisfaction and RS, p=0.008 and p=0.001). the obtained data suggest that rock style is related to the greater use of vocal strain and that this strain does not necessarily impose a negative impression to the voice, but corresponds to a common interpretative factor related to this style of music.

  7. Design and performance of a large vocabulary discrete word recognition system. Volume 1: Technical report. [real time computer technique for voice data processing

    NASA Technical Reports Server (NTRS)

    1973-01-01

    The development, construction, and test of a 100-word vocabulary near real time word recognition system are reported. Included are reasonable replacement of any one or all 100 words in the vocabulary, rapid learning of a new speaker, storage and retrieval of training sets, verbal or manual single word deletion, continuous adaptation with verbal or manual error correction, on-line verification of vocabulary as spoken, system modes selectable via verification display keyboard, relationship of classified word to neighboring word, and a versatile input/output interface to accommodate a variety of applications.

  8. Enhancement of temporal periodicity cues in cochlear implants: Effects on prosodic perception and vowel identification

    NASA Astrophysics Data System (ADS)

    Green, Tim; Faulkner, Andrew; Rosen, Stuart; Macherey, Olivier

    2005-07-01

    Standard continuous interleaved sampling processing, and a modified processing strategy designed to enhance temporal cues to voice pitch, were compared on tests of intonation perception, and vowel perception, both in implant users and in acoustic simulations. In standard processing, 400 Hz low-pass envelopes modulated either pulse trains (implant users) or noise carriers (simulations). In the modified strategy, slow-rate envelope modulations, which convey dynamic spectral variation crucial for speech understanding, were extracted by low-pass filtering (32 Hz). In addition, during voiced speech, higher-rate temporal modulation in each channel was provided by 100% amplitude-modulation by a sawtooth-like wave form whose periodicity followed the fundamental frequency (F0) of the input. Channel levels were determined by the product of the lower- and higher-rate modulation components. Both in acoustic simulations and in implant users, the ability to use intonation information to identify sentences as question or statement was significantly better with modified processing. However, while there was no difference in vowel recognition in the acoustic simulation, implant users performed worse with modified processing both in vowel recognition and in formant frequency discrimination. It appears that, while enhancing pitch perception, modified processing harmed the transmission of spectral information.

  9. Space Shuttle Orbiter audio subsystem. [to communication and tracking system

    NASA Technical Reports Server (NTRS)

    Stewart, C. H.

    1978-01-01

    The selection of the audio multiplex control configuration for the Space Shuttle Orbiter audio subsystem is discussed and special attention is given to the evaluation criteria of cost, weight and complexity. The specifications and design of the subsystem are described and detail is given to configurations of the audio terminal and audio central control unit (ATU, ACCU). The audio input from the ACCU, at a signal level of -12.2 to 14.8 dBV, nominal range, at 1 kHz, was found to have balanced source impedance and a balanced local impedance of 6000 + or - 600 ohms at 1 kHz, dc isolated. The Lyndon B. Johnson Space Center (JSC) electroacoustic test laboratory, an audio engineering facility consisting of a collection of acoustic test chambers, analyzed problems of speaker and headset performance, multiplexed control data coupled with audio channels, and the Orbiter cabin acoustic effects on the operational performance of voice communications. This system allows technical management and project engineering to address key constraining issues, such as identifying design deficiencies of the headset interface unit and the assessment of the Orbiter cabin performance of voice communications, which affect the subsystem development.

  10. Speaking Activities and Reading.

    ERIC Educational Resources Information Center

    Ediger, Marlow

    2000-01-01

    Notes that each pupil needs to receive guidance and assistance to achieve as optimally as possible in oral communication. Discusses critical listening to the spoken voice, using puppets, using role play activities, committees in the classroom, giving oral reports, oral reading to classmates, giving and following directions, extemporaneous…

  11. Co-Variation of Tonality in the Music and Speech of Different Cultures

    PubMed Central

    Han, Shui' er; Sundararajan, Janani; Bowling, Daniel Liu; Lake, Jessica; Purves, Dale

    2011-01-01

    Whereas the use of discrete pitch intervals is characteristic of most musical traditions, the size of the intervals and the way in which they are used is culturally specific. Here we examine the hypothesis that these differences arise because of a link between the tonal characteristics of a culture's music and its speech. We tested this idea by comparing pitch intervals in the traditional music of three tone language cultures (Chinese, Thai and Vietnamese) and three non-tone language cultures (American, French and German) with pitch intervals between voiced speech segments. Changes in pitch direction occur more frequently and pitch intervals are larger in the music of tone compared to non-tone language cultures. More frequent changes in pitch direction and larger pitch intervals are also apparent in the speech of tone compared to non-tone language cultures. These observations suggest that the different tonal preferences apparent in music across cultures are closely related to the differences in the tonal characteristics of voiced speech. PMID:21637716

  12. Regulation leads to increases in riparian vegetation, but not direct allochthonous inputs, along the Colorado River in Grand Canyon, Arizona

    USGS Publications Warehouse

    Kennedy, T.A.; Ralston, B.E.

    2012-01-01

    Dams and associated river regulation have led to the expansion of riparian vegetation, especially nonnative species, along downstream ecosystems. Nonnative saltcedar is one of the dominant riparian plants along virtually every major river system in the arid western United States, but allochthonous inputs have never been quantified along a segment of a large river that is dominated by saltcedar. We developed a novel method for estimating direct allochthonous inputs along the 387km-long reach of the Colorado River downstream of Glen Canyon Dam that utilized a GIS vegetation map developed from aerial photographs, empirical and literature-derived litter production data for the dominant vegetation types, and virtual shorelines of annual peak discharge (566m 3s -1 stage elevation). Using this method, we estimate that direct allochthonous inputs from riparian vegetation for the entire reach studied total 186metric tonsyear -1, which represents mean inputs of 470gAFDMm -1year -1 of shoreline or 5.17gAFDMm -2year -1 of river surface. These values are comparable to allochthonous inputs for other large rivers and systems that also have sparse riparian vegetation. Nonnative saltcedar represents a significant component of annual allochthonous inputs (36% of total direct inputs) in the Colorado River. We also estimated direct allochthonous inputs for 46.8km of the Colorado River prior to closure of Glen Canyon Dam using a vegetation map that was developed from historical photographs. Regulation has led to significant increases in riparian vegetation (270-319% increase in cover, depending on stage elevation), but annual allochthonous inputs appear unaffected by regulation because of the lower flood peaks on the post-dam river. Published in 2010 by John Wiley & Sons, Ltd.

  13. It's not what you hear, it's the way you think about it: appraisals as determinants of affect and behaviour in voice hearers.

    PubMed

    Peters, E R; Williams, S L; Cooke, M A; Kuipers, E

    2012-07-01

    Previous studies have suggested that beliefs about voices mediate the relationship between actual voice experience and behavioural and affective response. We investigated beliefs about voice power (omnipotence), voice intent (malevolence/benevolence) and emotional and behavioural response (resistance/engagement) using the Beliefs About Voices Questionnaire - Revised (BAVQ-R) in 46 voice hearers. Distress was assessed using a wide range of measures: voice-related distress, depression, anxiety, self-esteem and suicidal ideation. Voice topography was assessed using measures of voice severity, frequency and intensity. We predicted that beliefs about voices would show a stronger association with distress than voice topography. Omnipotence had the strongest associations with all measures of distress included in the study whereas malevolence was related to resistance, and benevolence to engagement. As predicted, voice severity, frequency and intensity were not related to distress once beliefs were accounted for. These results concur with previous findings that beliefs about voice power are key determinants of distress in voice hearers, and should be targeted specifically in psychological interventions.

  14. Updating signal typing in voice: addition of type 4 signals.

    PubMed

    Sprecher, Alicia; Olszewski, Aleksandra; Jiang, Jack J; Zhang, Yu

    2010-06-01

    The addition of a fourth type of voice to Titze's voice classification scheme is proposed. This fourth voice type is characterized by primarily stochastic noise behavior and is therefore unsuitable for both perturbation and correlation dimension analysis. Forty voice samples were classified into the proposed four types using narrowband spectrograms. Acoustic, perceptual, and correlation dimension analyses were completed for all voice samples. Perturbation measures tended to increase with voice type. Based on reliability cutoffs, the type 1 and type 2 voices were considered suitable for perturbation analysis. Measures of unreliability were higher for type 3 and 4 voices. Correlation dimension analyses increased significantly with signal type as indicated by a one-way analysis of variance. Notably, correlation dimension analysis could not quantify the type 4 voices. The proposed fourth voice type represents a subset of voices dominated by noise behavior. Current measures capable of evaluating type 4 voices provide only qualitative data (spectrograms, perceptual analysis, and an infinite correlation dimension). Type 4 voices are highly complex and the development of objective measures capable of analyzing these voices remains a topic of future investigation.

  15. Mechanics of human voice production and control

    PubMed Central

    Zhang, Zhaoyan

    2016-01-01

    As the primary means of communication, voice plays an important role in daily life. Voice also conveys personal information such as social status, personal traits, and the emotional state of the speaker. Mechanically, voice production involves complex fluid-structure interaction within the glottis and its control by laryngeal muscle activation. An important goal of voice research is to establish a causal theory linking voice physiology and biomechanics to how speakers use and control voice to communicate meaning and personal information. Establishing such a causal theory has important implications for clinical voice management, voice training, and many speech technology applications. This paper provides a review of voice physiology and biomechanics, the physics of vocal fold vibration and sound production, and laryngeal muscular control of the fundamental frequency of voice, vocal intensity, and voice quality. Current efforts to develop mechanical and computational models of voice production are also critically reviewed. Finally, issues and future challenges in developing a causal theory of voice production and perception are discussed. PMID:27794319

  16. Mechanics of human voice production and control.

    PubMed

    Zhang, Zhaoyan

    2016-10-01

    As the primary means of communication, voice plays an important role in daily life. Voice also conveys personal information such as social status, personal traits, and the emotional state of the speaker. Mechanically, voice production involves complex fluid-structure interaction within the glottis and its control by laryngeal muscle activation. An important goal of voice research is to establish a causal theory linking voice physiology and biomechanics to how speakers use and control voice to communicate meaning and personal information. Establishing such a causal theory has important implications for clinical voice management, voice training, and many speech technology applications. This paper provides a review of voice physiology and biomechanics, the physics of vocal fold vibration and sound production, and laryngeal muscular control of the fundamental frequency of voice, vocal intensity, and voice quality. Current efforts to develop mechanical and computational models of voice production are also critically reviewed. Finally, issues and future challenges in developing a causal theory of voice production and perception are discussed.

  17. Voice care knowledge among clinicians and people with healthy voices or dysphonia.

    PubMed

    Fletcher, Helen M; Drinnan, Michael J; Carding, Paul N

    2007-01-01

    An important clinical component in the prevention and treatment of voice disorders is voice care and hygiene. Research in voice care knowledge has mainly focussed on specific groups of professional voice users with limited reporting on the tool and evidence base used. In this study, a questionnaire to measure voice care knowledge was developed based on "best evidence." The questionnaire was validated by measuring specialist voice clinicians' agreement. Preliminary data are then presented using the voice care knowledge questionnaire with 17 subjects with nonorganic dysphonia and 17 with healthy voices. There was high (89%) agreement among the clinicians. There was a highly significant difference between the dysphonic and the healthy group scores (P = 0.00005). Furthermore, the dysphonic subjects (63% agreement) presented with less voice care knowledge than the subjects with healthy voices (72% agreement). The questionnaire provides a useful and valid tool to investigate voice care knowledge. The findings have implications for clinical intervention, voice therapy, and health prevention.

  18. Input and Uptake at 7 Months Predicts Toddler Vocabulary: The Role of Child-Directed Speech and Infant Processing Skills in Language Development

    ERIC Educational Resources Information Center

    Newman, Rochelle S.; Rowe, Meredith L.; Ratner, Nan Bernstein

    2016-01-01

    Both the input directed to the child, and the child's ability to process that input, are likely to impact the child's language acquisition. We explore how these factors inter-relate by tracking the relationships among: (a) lexical properties of maternal child-directed speech to prelinguistic (7-month-old) infants (N = 121); (b) these infants'…

  19. Mothers Consistently Alter Their Unique Vocal Fingerprints When Communicating with Infants.

    PubMed

    Piazza, Elise A; Iordan, Marius Cătălin; Lew-Williams, Casey

    2017-10-23

    The voice is the most direct link we have to others' minds, allowing us to communicate using a rich variety of speech cues [1, 2]. This link is particularly critical early in life as parents draw infants into the structure of their environment using infant-directed speech (IDS), a communicative code with unique pitch and rhythmic characteristics relative to adult-directed speech (ADS) [3, 4]. To begin breaking into language, infants must discern subtle statistical differences about people and voices in order to direct their attention toward the most relevant signals. Here, we uncover a new defining feature of IDS: mothers significantly alter statistical properties of vocal timbre when speaking to their infants. Timbre, the tone color or unique quality of a sound, is a spectral fingerprint that helps us instantly identify and classify sound sources, such as individual people and musical instruments [5-7]. We recorded 24 mothers' naturalistic speech while they interacted with their infants and with adult experimenters in their native language. Half of the participants were English speakers, and half were not. Using a support vector machine classifier, we found that mothers consistently shifted their timbre between ADS and IDS. Importantly, this shift was similar across languages, suggesting that such alterations of timbre may be universal. These findings have theoretical implications for understanding how infants tune in to their local communicative environments. Moreover, our classification algorithm for identifying infant-directed timbre has direct translational implications for speech recognition technology. Copyright © 2017 Elsevier Ltd. All rights reserved.

  20. Quantitative analysis of professionally trained versus untrained voices.

    PubMed

    Siupsinskiene, Nora

    2003-01-01

    The aim of this study was to compare healthy trained and untrained voices as well as healthy and dysphonic trained voices in adults using combined voice range profile and aerodynamic tests, to define the normal range limiting values of quantitative voice parameters and to select the most informative quantitative voice parameters for separation between healthy and dysphonic trained voices. Three groups of persons were evaluated. One hundred eighty six healthy volunteers were divided into two groups according to voice training: non-professional speakers group consisted of 106 untrained voices persons (36 males and 70 females) and professional speakers group--of 80 trained voices persons (21 males and 59 females). Clinical group consisted of 103 dysphonic professional speakers (23 males and 80 females) with various voice disorders. Eighteen quantitative voice parameters from combined voice range profile (VRP) test were analyzed: 8 of voice range profile, 8 of speaking voice, overall vocal dysfunction degree and coefficient of sound, and aerodynamic maximum phonation time. Analysis showed that healthy professional speakers demonstrated expanded vocal abilities in comparison to healthy non-professional speakers. Quantitative voice range profile parameters- pitch range, high frequency limit, area of high frequencies and coefficient of sound differed significantly between healthy professional and non-professional voices, and were more informative than speaking voice or aerodynamic parameters in showing the voice training. Logistic stepwise regression revealed that VRP area in high frequencies was sufficient to discriminate between healthy and dysphonic professional speakers for male subjects (overall discrimination accuracy--81.8%) and combination of three quantitative parameters (VRP high frequency limit, maximum voice intensity and slope of speaking curve) for female subjects (overall model discrimination accuracy--75.4%). We concluded that quantitative voice assessment with selected parameters might be useful for evaluation of voice education for healthy professional speakers as well as for detection of vocal dysfunction and evaluation of rehabilitation effect in dysphonic professionals.

  1. The Voice as Computer Interface: A Look at Tomorrow's Technologies.

    ERIC Educational Resources Information Center

    Lange, Holley R.

    1991-01-01

    Discussion of voice as the communications device for computer-human interaction focuses on voice recognition systems for use within a library environment. Voice technologies are described, including voice response and voice recognition; examples of voice systems in use in libraries are examined; and further possibilities, including use with…

  2. [The voice of the singer in the phonetogram].

    PubMed

    Klingholz, F

    1989-01-01

    Phonetograms were subdivided into areas approximating voice registers. By means of an analytical description of the areas, parameters could be established for a differentiation of voice categories and efficiency. The evaluation of 21 untrained and 34 trained voices showed a significant difference between the two groups. Male singers demonstrated more efficiency in the head and chest registers than male non-singers; female singers showed a stronger efficiency only in the head voice in comparison with their non-singer counterparts. Proceeding from voice sound alone, voices are often misclassified regarding the voice categories, and voice problems arise. Moreover, enhanced training of only chest or head voice function results in functional disorders in the singing voice. Such cases can be demonstrated by means of phonetograms.

  3. Variations in Intensity, Fundamental Frequency, and Voicing for Teachers in Occupational Versus Non-Occupational Settings

    PubMed Central

    Hunter, Eric J.; Titze, Ingo R.

    2012-01-01

    Purpose This study creates a more concise picture of the vocal demands placed on teachers by comparing occupational voice use with non-occupational voice use. Methods The National Center for Voice and Speech voice dosimetry databank was used to calculate voicing percentage per hour, as well as average dB SPL and F0. Occupational voice use (9am-3 PM, weekdays) and non-occupational voice use (4 PM-10 PM, weekends) were compared (57 teachers, two weeks each). Results Five key findings were uncovered: [1] similar to previous studies, occupational voicing percentage per hour is more than twice that of non-occupational; [2] teachers experienced a wide range of occupational voicing percentages per hour (30±11%/hr); [3] average occupational voice was about 1 dB SPL louder than the non-occupational voice and remained constant throughout the day; [4] occupational voice exhibited an increased pitch and trended upward throughout the day; [5] some apparent gender differences were shown. Conclusions Data regarding voicing percentages, F0 and dB SPL provide critical insight into teachers’ vocal health. Further, because non-occupational voice use is added to an already overloaded voice, it may add key insights into recovery patterns, and should be the focus of future studies. PMID:20689046

  4. Temporal voice areas exist in autism spectrum disorder but are dysfunctional for voice identity recognition

    PubMed Central

    Borowiak, Kamila; von Kriegstein, Katharina

    2016-01-01

    The ability to recognise the identity of others is a key requirement for successful communication. Brain regions that respond selectively to voices exist in humans from early infancy on. Currently, it is unclear whether dysfunction of these voice-sensitive regions can explain voice identity recognition impairments. Here, we used two independent functional magnetic resonance imaging studies to investigate voice processing in a population that has been reported to have no voice-sensitive regions: autism spectrum disorder (ASD). Our results refute the earlier report that individuals with ASD have no responses in voice-sensitive regions: Passive listening to vocal, compared to non-vocal, sounds elicited typical responses in voice-sensitive regions in the high-functioning ASD group and controls. In contrast, the ASD group had a dysfunction in voice-sensitive regions during voice identity but not speech recognition in the right posterior superior temporal sulcus/gyrus (STS/STG)—a region implicated in processing complex spectrotemporal voice features and unfamiliar voices. The right anterior STS/STG correlated with voice identity recognition performance in controls but not in the ASD group. The findings suggest that right STS/STG dysfunction is critical for explaining voice recognition impairments in high-functioning ASD and show that ASD is not characterised by a general lack of voice-sensitive responses. PMID:27369067

  5. Provision of surgical voice restoration in England: questionnaire survey of speech and language therapists.

    PubMed

    Bradley, P J; Counter, P; Hurren, A; Cocks, H C

    2013-08-01

    To conduct a questionnaire survey of speech and language therapists providing and managing surgical voice restoration in England. National Health Service Trusts registering more than 10 new laryngeal cancer patients during any one year, from November 2009 to October 2010, were identified, and a list of speech and language therapists compiled. A questionnaire was developed, peer reviewed and revised. The final questionnaire was e-mailed with a covering letter to 82 units. Eighty-two questionnaires were distributed and 72 were returned and analysed, giving a response rate of 87.8 per cent. Forty-four per cent (38/59) of the units performed more than 10 laryngectomies per year. An in-hours surgical voice restoration service was provided by speech and language therapists in 45.8 per cent (33/72) and assisted by nurses in 34.7 per cent (25/72). An out of hours service was provided directly by ENT staff in 35.5 per cent (21/59). Eighty-eight per cent (63/72) of units reported less than 10 (emergency) out of hours calls per month. Surgical voice restoration service provision varies within and between cancer networks. There is a need for a national management and care protocol, an educational programme for out of hours service providers, and a review of current speech and language therapist staffing levels in England.

  6. Satellite voice broadcast system study, volume 2

    NASA Technical Reports Server (NTRS)

    Horstein, M.

    1985-01-01

    This study investigates the feasibility of providing Voice of America (VOA) broadcasts by satellite relay, rather than via terrestrial relay stations. Satellite voice broadcast systems are described for three different frequency bands: HF (26 MHz), VHF (68 MHz), and L-band (1.5 GHz). The geographical areas of interest at HF and L-band include all major land masses worldwide with the exception of the U.S., Canada, and Australia. Geostationary satellite configurations are considered for both frequency bands. In addition, a system of subsynchronous, circular satellites with an orbit period of 8 hours is developed for the HF band. VHF broadcasts, which are confined to the Soviet Union, are provied by a system of Molniya satellites. Satellites intended for HF or VHF broadcastinbg are extremely large and heavy. Satellite designs presented here are limited in size and weight to the capability of the STS/Centaur launch vehicle combination. Even so, at HF it would take 47 geostationary satellites or 20 satellites in 8-hour orbits to fully satisfy the voice-channel requirements of the broadcast schedule provided by VOA. On the other hand, three Molniya satellites suffice for the geographically restricted schedule at VHF. At L-band, only four geostationary satellites are needed to meet the requirements of the complete broadcast schedule. Moreover, these satellites are comparable in size and weight to current satellites designed for direct broadcast of video program material.

  7. Integrated Software Systems for Crew Management During Extravehicular Activity in Planetary Terrain Exploration

    NASA Technical Reports Server (NTRS)

    Kuznetz, Lawrence; Nguen, Dan; Jones, Jeffrey; Lee, Pascal; Merrell, Ronald; Rafiq, Azhar

    2008-01-01

    Initial planetary explorations with the Apollo program had a veritable ground support army monitoring the safety and health of the 12 astronauts who performed lunar surface extravehicular activities (EVAs). Given the distances involved, this will not be possible on Mars. A spacesuit for Mars must be smart enough to replace that army. The next generation suits can do so using 2 software systems serving as virtual companions, LEGACI (Life support, Exploration Guidance Algorithm and Consumable Interrogator) and VIOLET (Voice Initiated Operator for Life support and Exploration Tracking). The system presented in this study integrates data inputs from a suite of sensors into the MIII suit s communications, avionics and informatics hardware for distribution to remote managers and data analysis. If successful, the system has application not only for Mars but for nearer term missions to the Moon, and the next generation suits used on ISS as well. Field tests are conducted to assess capabilities for next generation spacesuits at Johnson Space Center (JSC) as well as the Mars and Lunar analog (Devon Island, Canada). LEGACI integrates data inputs from a suite of noninvasive biosensors in the suit and the astronaut (heart rate, suit inlet/outlet lcg temperature and flowrate, suit outlet gas and dewpoint temperature, pCO2, suit O2 pressure, state vector (accelerometry) and others). In the Integrated Walkback Suit Tests held at NASA-JSC and the HMP tests at Devon Island, communication and informatics capabilities were tested (including routing by satellite from the suit at Devon Island to JSC in Houston via secure servers at VCU in Richmond, VA). Results. The input from all the sensors enable LEGACI to compute multiple independent assessments of metabolic rate, from which a "best" met rate is chosen based on statistical methods. This rate can compute detailed information about the suit, crew and EVA performance using test-derived algorithms. VIOLET gives LEGACI voice activation capability, allowing the crew to query the suit, and receive feedback and alerts that will lead to corrective action. LEGACI and VIOLET can also automatically control the astronaut's cooling and consumable use rate without crew input if desired. These findings suggest that non-invasive physiological and environmental sensors supported with data analysis can allow for more effective management of mission task performance during EVA. Integrated remote and local view of data metrics allow crewmember to receive real time feedback in synch with mission control in preventing performance shortcomings for EVA in exploration missions.

  8. The prevalence of voice disorders in 911 emergency telecommunicators.

    PubMed

    Johns-Fiedler, Heidi; van Mersbergen, Miriam

    2015-05-01

    Emergency 911 dispatchers or telecommunicators have been cited as occupational voice users who could be at risk for voice disorders. To test the theoretical assumption that the 911 emergency telecommunicators (911ETCs) are exposed to risk for voice disorders because of their heavy vocal load, this study assessed the prevalence of voice complaints in 911ETCs. A cross-sectional survey was sent to two large national organizations for 911ETCs with 71 complete responses providing information about voice health, voice complaints, and work load. Although 911ETCs have a higher rate of reported voice symptoms and score higher on the Voice Handicap Index-10 than the general public, they have a voice disorder diagnosis prevalence that mirrors the prevalence of the general population. The 911ETCs may be underserved in the voice community and would benefit from education on vocal health and treatments for voice complaints. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  9. Integrating cues of social interest and voice pitch in men's preferences for women's voices.

    PubMed

    Jones, Benedict C; Feinberg, David R; Debruine, Lisa M; Little, Anthony C; Vukovic, Jovana

    2008-04-23

    Most previous studies of vocal attractiveness have focused on preferences for physical characteristics of voices such as pitch. Here we examine the content of vocalizations in interaction with such physical traits, finding that vocal cues of social interest modulate the strength of men's preferences for raised pitch in women's voices. Men showed stronger preferences for raised pitch when judging the voices of women who appeared interested in the listener than when judging the voices of women who appeared relatively disinterested in the listener. These findings show that voice preferences are not determined solely by physical properties of voices and that men integrate information about voice pitch and the degree of social interest expressed by women when forming voice preferences. Women's preferences for raised pitch in women's voices were not modulated by cues of social interest, suggesting that the integration of cues of social interest and voice pitch when men judge the attractiveness of women's voices may reflect adaptations that promote efficient allocation of men's mating effort.

  10. I like my voice better: self-enhancement bias in perceptions of voice attractiveness.

    PubMed

    Hughes, Susan M; Harrison, Marissa A

    2013-01-01

    Previous research shows that the human voice can communicate a wealth of nonsemantic information; preferences for voices can predict health, fertility, and genetic quality of the speaker, and people often use voice attractiveness, in particular, to make these assessments of others. But it is not known what we think of the attractiveness of our own voices as others hear them. In this study eighty men and women rated the attractiveness of an array of voice recordings of different individuals and were not told that their own recorded voices were included in the presentation. Results showed that participants rated their own voices as sounding more attractive than others had rated their voices, and participants also rated their own voices as sounding more attractive than they had rated the voices of others. These findings suggest that people may engage in vocal implicit egotism, a form of self-enhancement.

  11. Factors Affecting the Implementation of Argument in the Elementary Science Classroom. A Longitudinal Case Study

    NASA Astrophysics Data System (ADS)

    Martin, Anita M.; Hand, Brian

    2009-01-01

    This longitudinal case study describes the factors that affect an experienced teacher’s attempt to shift her pedagogical practices in order to implement embedded elements of argument into her science classroom. Research data was accumulated over 2 years through video recordings of science classes. The Reformed Teacher Observation Protocol (RTOP) is an instrument designed to quantify changes in classroom environments as related to reform as defined by the National Research Council ( National science education standards. Washington, DC: National Academy Press, 1996b) and the National Research Council ( Fulfilling the promise: Biology education in the nation’s schools, Washington, DC: National Academy Press, 1990) and was used to analyze videotaped science lessons. Analysis of the data shows that there was a significant shift in the areas of teacher questioning, and student voice. Several levels of subsequent analysis were completed related to teacher questioning and student voice. The data suggests a relationship between these areas and the implementation of scientific argument. Results indicate that the teacher moved from a traditional, teacher-centered, didactic teaching style to instructional practices that allowed the focus and direction of the lesson to be affected by student voice. This was accomplished by a change in teacher questioning that included a shift from factual recall to more divergent questioning patterns allowing for increased student voice. As student voice increased, students began to investigate ideas, make statements or claims and to support these claims with strong evidence. Finally, students were observed refuting claims in the form of rebuttals. This study informs professional development related to experienced teachers in that it highlights pedagogical issues involved in implementing embedded elements of argument in the elementary classroom.

  12. Relationship between patient-perceived vocal handicap and clinician-rated level of vocal dysfunction.

    PubMed

    Childs, Lesley F; Bielinski, Clifford; Toles, Laura; Hamilton, Amy; Deane, Janis; Mau, Ted

    2015-01-01

    The relationship between patient-reported vocal handicap and clinician-rated measures of vocal dysfunction is not understood. This study aimed to determine if a correlation exists between the Voice Handicap Index-10 (VHI-10) and the Voice Functional Communication Measure rating in the National Outcomes Measurement System (NOMS). Retrospective case series. Four hundred and nine voice evaluations over 12 months at a tertiary voice center were reviewed. The VHI-10 and NOMS scores, diagnoses, and potential comorbid factors were collected and analyzed. For the study population as a whole, there was a moderate negative correlation between the NOMS rating and the VHI-10 (Pearson r = -0.57). However, for a given NOMS level, there could be considerable spread in the VHI-10. In addition, as the NOMS decreased stepwise below level 4, there was a corresponding increase in the VHI-10. However, a similar trend in VHI-10 was not observed for NOMS above level 4, indicating the NOMS versus VHI-10 correlation was not linear. Among diagnostic groups, the strongest correlation was found for subjects with functional dysphonia. The NOMS versus VHI-10 correlation was not affected by gender or the coexistence of a psychiatric diagnosis. A simple relationship between VHI-10 and NOMS rating does not exist. Patients with mild vocal dysfunction have a less direct relationship between their NOMS ratings and the VHI-10. These findings provide insight into the interpretation of patient-perceived and clinician-rated measures of vocal function and may allow for better management of expectations and patient counseling in the treatment of voice disorders. © 2014 The American Laryngological, Rhinological and Otological Society, Inc.

  13. Project planning, training, measurement and sustainment: the successful implementation of voice recognition.

    PubMed

    Antiles, S; Couris, J; Schweitzer, A; Rosenthal, D; Da Silva, R Q

    2000-01-01

    Computerized voice recognition systems (VR) can reduce costs and enhance service. The capital outlay required for conversion to a VR system is significant; therefore, it is incumbent on radiology departments to provide cost and service justifications to administrators. Massachusetts General Hospital (MGH) in Boston implemented VR over a two-year period and achieved annual savings of $530,000 and a 50% decrease in report throughput. Those accomplishments required solid planning and implementation strategies, training and sustainment programs. This article walks through the process, step by step, in the hope of providing a tool set for future implementations. Because VR has dramatic implications for workflow, a solid operational plan is needed when assessing vendors and planning for implementation. The goals for implementation should be to minimize operational disruptions and capitalize on efficiencies of the technology. Senior leadership--the department chair or vice-chair--must select the goals to be accomplished and oversee, manage and direct the VR initiative. The importance of this point cannot be overstated, since implementation will require behavior changes from radiologists and others who may not perceive any personal benefits. Training is the pivotal factor affecting the success of voice recognition, and practice is the only way for radiologists to enhance their skills. Through practice, radiologists will discover shortcuts, and their speed and comfort will improve. Measurement and data analysis are critical to changing and improving the voice recognition application and are vital to decision-making. Some of the issues about which valuable date can be collected are technical and educational problems, VR penetration, report turnaround time and annual cost savings. Sustained effort is indispensable to the maintenance of voice recognition. Finally, all efforts made and gains achieved may prove to be futile without ongoing sustainment of the system through retraining, education and technical support.

  14. Computerized Analysis of Acoustic Characteristics of Patients with Internal Nasal Valve Collapse Before and After Functional Rhinoplasty

    PubMed Central

    Rezaei, Fariba; Omrani, Mohammad Reza; Abnavi, Fateme; Mojiri, Fariba; Golabbakhsh, Marzieh; Barati, Sohrab; Mahaki, Behzad

    2015-01-01

    Acoustic analysis of sounds produced during speech provides significant information about the physiology of larynx and vocal tract. The analysis of voice power spectrum is a fundamental sensitive method of acoustic assessment that provides valuable information about the voice source and characteristics of vocal tract resonance cavities. The changes in long-term average spectrum (LTAS) spectral tilt and harmony to noise ratio (HNR) were analyzed to assess the voice quality before and after functional rhinoplasty in patients with internal nasal valve collapse. Before and 3 months after functional rhinoplasty, 12 participants were evaluated and HNR and LTAS spectral tilt in /a/ and /i/ vowels were estimated. It was seen that an increase in HNR and a decrease in LTAS spectral tilt existed after surgery. Mean LTAS spectral tilt in vowel /a/ decreased from 2.37 ± 1.04 to 2.28 ± 1.17 (P = 0.388), and it was decreased from 4.16 ± 1.65 to 2.73 ± 0.69 in vowel /i/ (P = 0.008). Mean HNR in the vowel /a/ increased from 20.71 ± 3.93 to 25.06 ± 2.67 (P = 0.002), and it was increased from 21.28 ± 4.11 to 25.26 ± 3.94 in vowel /i/ (P = 0.002). Modification of the vocal tract caused the vocal cords to close sufficiently, and this showed that although rhinoplasty did not affect the larynx directly, it changes the structure of the vocal tract and consequently the resonance of voice production. The aim of this study was to investigate the changes in voice parameters after functional rhinoplasty in patients with internal nasal valve collapse by computerized analysis of acoustic characteristics. PMID:26955564

  15. The Effects of the Menstrual Cycle on Vibratory Characteristics of the Vocal Folds Investigated With High-Speed Digital Imaging.

    PubMed

    Kunduk, Melda; Vansant, Mathew B; Ikuma, Takeshi; McWhorter, Andrew

    2017-03-01

    This study investigated the effect of menstrual cycle on vocal fold vibratory characteristics in young women using high-speed digital imaging. This study examined the menstrual phase effect on five objective high-speed imaging parameters and two self-rated perceptual parameters. The effects of oral birth control use were also investigated. Thirteen subjects with no prior voice complaints were included in this study. All data were collected at three different time periods (premenses, postmenses, ovulation) over the course of one menstrual cycle. For five of the 13 subjects, data were collected for two consecutive cycles. Six of 13 subjects were oral birth control users. From high-speed imaging data, five objective parameters were computed: fundamental frequency, fundamental frequency deviation, harmonics-to-noise ratio, harmonic richness factor, and ratio of first and second harmonics. They were supplemented by two self-rated parameters: Reflux Severity Index and perceptual voice quality rating. Analysis included mixed model linear analysis with repeated measures. Results indicated no significant main effects for menstrual phase, between-cycle, or birth control use in the analysis for mean fundamental frequency, fundamental frequency deviation, harmonics-to-noise ratio, harmonic richness factor, first and second harmonics, Reflux Severity Index, and perceptual voice quality rating. Additionally, there were no interaction effects. Hormone fluctuations observed across the menstrual cycle do not appear to have direct effect on vocal fold vibratory characteristics in young women with no voice concerns. Birth control use, on the other hand, may have influence on spectral richness of vocal fold vibration. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  16. Developing technologies for bioacoustic vocal profiling as a viable component of integrative medical diagnostics and treatment

    NASA Astrophysics Data System (ADS)

    Edwards, Sharry K.

    2005-04-01

    Over the past 20+ years the pioneering field of Human Bioacoustics, which includes voice spectral analysis, has begun to model the frequencies and architecture of human vocalizations to identify the innate mathematical templates found within the various system of the human body. Using the idea that the voice is a holographic representation of health and wellness, these non-invasive techniques are being advanced to the extent that a computerized Vocal Profile, using a system of Frequency Equivalents, can be used to accurately quantify, organize, interpret, define, and extrapolate biometric information from the human voice. This information, in turn, provides the opportunity to predict, direct, and maintain intrinsic form and function. This novel approach has provided an accumulation of significant data but until recently has been without an efficient biological framework of reference. The emerging Mathematical Model being assembled through Human Bioacoustic research likely has the potential to allow Vocal Profiling to be used to predict and monitor health issues from the very first cries of a newborn through the frequency foundations of disease and aging.

  17. Distinguishing between forensic science and forensic pseudoscience: testing of validity and reliability, and approaches to forensic voice comparison.

    PubMed

    Morrison, Geoffrey Stewart

    2014-05-01

    In this paper it is argued that one should not attempt to directly assess whether a forensic analysis technique is scientifically acceptable. Rather one should first specify what one considers to be appropriate principles governing acceptable practice, then consider any particular approach in light of those principles. This paper focuses on one principle: the validity and reliability of an approach should be empirically tested under conditions reflecting those of the case under investigation using test data drawn from the relevant population. Versions of this principle have been key elements in several reports on forensic science, including forensic voice comparison, published over the last four-and-a-half decades. The aural-spectrographic approach to forensic voice comparison (also known as "voiceprint" or "voicegram" examination) and the currently widely practiced auditory-acoustic-phonetic approach are considered in light of this principle (these two approaches do not appear to be mutually exclusive). Approaches based on data, quantitative measurements, and statistical models are also considered in light of this principle. © 2013.

  18. Resonance strategies used in Bulgarian women's singing style: a pilot study.

    PubMed

    Henrich, Nathalie; Kiek, Mara; Smith, John; Wolfe, Joe

    2007-01-01

    Are the characteristic timbre and loudness of Bulgarian women's singing related to tuning of resonances of the vocal tract? We studied an Australian female singer, who practises and teaches Bulgarian singing technique. Two different vocal qualities of this style were studied. The louder teshka is characterized by a sonorous voice production. The less loud leka has a smoother timbre that is closer to that of the head voice register. Six vowels in each of teshka, leka and the subject's 'normal' (i.e. Western rather than Bulgarian) style were studied. The acoustic resonances of the singer's vocal tract were measured directly during singing by injecting a synthesized, broad-band acoustic current. This singer does not use resonance tuning consistently in her classical Western style. However, in both teshka and leka, she tunes the first tract resonance close to the second harmonic of the voice for most vowels. This tuning boosts the power output in the radiation field for that harmonic. This tuning also contributes to the very strong second harmonic which is a characteristic of the timbre identified as the Bulgarian style.

  19. Hands-free human-machine interaction with voice

    NASA Astrophysics Data System (ADS)

    Juang, B. H.

    2004-05-01

    Voice is natural communication interface between a human and a machine. The machine, when placed in today's communication networks, may be configured to provide automation to save substantial operating cost, as demonstrated in AT&T's VRCP (Voice Recognition Call Processing), or to facilitate intelligent services, such as virtual personal assistants, to enhance individual productivity. These intelligent services often need to be accessible anytime, anywhere (e.g., in cars when the user is in a hands-busy-eyes-busy situation or during meetings where constantly talking to a microphone is either undersirable or impossible), and thus call for advanced signal processing and automatic speech recognition techniques which support what we call ``hands-free'' human-machine communication. These techniques entail a broad spectrum of technical ideas, ranging from use of directional microphones and acoustic echo cancellatiion to robust speech recognition. In this talk, we highlight a number of key techniques that were developed for hands-free human-machine communication in the mid-1990s after Bell Labs became a unit of Lucent Technologies. A video clip will be played to demonstrate the accomplishement.

  20. Studies in automatic speech recognition and its application in aerospace

    NASA Astrophysics Data System (ADS)

    Taylor, Michael Robinson

    Human communication is characterized in terms of the spectral and temporal dimensions of speech waveforms. Electronic speech recognition strategies based on Dynamic Time Warping and Markov Model algorithms are described and typical digit recognition error rates are tabulated. The application of Direct Voice Input (DVI) as an interface between man and machine is explored within the context of civil and military aerospace programmes. Sources of physical and emotional stress affecting speech production within military high performance aircraft are identified. Experimental results are reported which quantify fundamental frequency and coarse temporal dimensions of male speech as a function of the vibration, linear acceleration and noise levels typical of aerospace environments; preliminary indications of acoustic phonetic variability reported by other researchers are summarized. Connected whole-word pattern recognition error rates are presented for digits spoken under controlled Gz sinusoidal whole-body vibration. Correlations are made between significant increases in recognition error rate and resonance of the abdomen-thorax and head subsystems of the body. The phenomenon of vibrato style speech produced under low frequency whole-body Gz vibration is also examined. Interactive DVI system architectures and avionic data bus integration concepts are outlined together with design procedures for the efficient development of pilot-vehicle command and control protocols.

  1. Diabetes Bingo: Research Prioritization with the Filipino Community

    PubMed Central

    Oculto, Tessie; Ramones, Emilyn; Caagbay, Cedric R

    2010-01-01

    This community-based participatory research, conducted in partnership between a European-American academic researcher and a professional group of Filipino nurses, aimed to determine the diabetes research priority for the Filipino community on the island of O‘ahu in Hawai‘i, and to evaluate the multi-voting technique to seek input from the community. The study design was a qualitative, cross-sectional interactive process consisting of an educational presentation followed by data collection from the audience. Ten community presentations about the impact of diabetes on the Filipino community were conducted by a Filipino nurse with participants (N = 265). Following the educational session, the participants selected priorities for research using a multi-vote technique developed as a Diabetes Bingo card. Community voting results identified prevention and a focus on adults as important priorities for research. Based on the results of the multi-voting, the research partners were able to come to consensus on a research priority area of prevention of type 2 diabetes in adults. Multi-voting using a Diabetes Bingo card, preceded by an educational presentation by a Filipino nurse, was a culturally competent community-based participatory research method that gave voice to the participants and direction to the research partners for future projects. The multi-voting technique was readily accepted and enjoyed by participants. PMID:21229487

  2. Self-organizing feature maps for dynamic control of radio resources in CDMA microcellular networks

    NASA Astrophysics Data System (ADS)

    Hortos, William S.

    1998-03-01

    The application of artificial neural networks to the channel assignment problem for cellular code-division multiple access (CDMA) cellular networks has previously been investigated. CDMA takes advantage of voice activity and spatial isolation because its capacity is only interference limited, unlike time-division multiple access (TDMA) and frequency-division multiple access (FDMA) where capacities are bandwidth-limited. Any reduction in interference in CDMA translates linearly into increased capacity. To satisfy the high demands for new services and improved connectivity for mobile communications, microcellular and picocellular systems are being introduced. For these systems, there is a need to develop robust and efficient management procedures for the allocation of power and spectrum to maximize radio capacity. Topology-conserving mappings play an important role in the biological processing of sensory inputs. The same principles underlying Kohonen's self-organizing feature maps (SOFMs) are applied to the adaptive control of radio resources to minimize interference, hence, maximize capacity in direct-sequence (DS) CDMA networks. The approach based on SOFMs is applied to some published examples of both theoretical and empirical models of DS/CDMA microcellular networks in metropolitan areas. The results of the approach for these examples are informally compared to the performance of algorithms, based on Hopfield- Tank neural networks and on genetic algorithms, for the channel assignment problem.

  3. Voice-on-Target: A New Approach to Tactical Networking and Unmanned Systems Control via the Voice Interface to the SA Environment

    DTIC Science & Technology

    2009-06-01

    Blackberry handheld) device. After each voice command activation, the medic provided voice comments to be recorded in Observer Notepad over Voice...vial (up-right corner of picture) upon voice activation from the medic’s Blackberry handheld. The NPS UAS which was controlled by voice commands...Voice Portal using a standard Blackberry handheld with a head set. The results demonstrated sufficient accuracy for controlling the tactical sensor

  4. Electroacoustic Performance of Direct-Input Hearing Aids with FM Amplification Systems.

    ERIC Educational Resources Information Center

    Thibodeau, Linda M.

    1990-01-01

    The electroacoustic performance of 18 direct-input and two inductive-coupling hearing aids was compared when operating with two different frequency modulation (FM) systems. The most significant differences occurred in full-on gain, equivalent-input noise, and frequency response, as opposed to high frequency average saturation sound pressure level…

  5. Cognitive Attachment Model of Voices: Evidence Base and Future Implications

    PubMed Central

    Berry, Katherine; Varese, Filippo; Bucci, Sandra

    2017-01-01

    There is a robust association between hearing voices and exposure to traumatic events. Identifying mediating mechanisms for this relationship is key to theories of voice hearing and the development of therapies for distressing voices. This paper outlines the Cognitive Attachment model of Voices (CAV), a theoretical model to understand the relationship between earlier interpersonal trauma and distressing voice hearing. The model builds on attachment theory and well-established cognitive models of voices and argues that attachment and dissociative processes are key psychological mechanisms that explain how trauma influences voice hearing. Following the presentation of the model, the paper will review the current state of evidence regarding the proposed mechanisms of vulnerability to voice hearing and maintenance of voice-related distress. This review will include evidence from studies supporting associations between dissociation and voices, followed by details of our own research supporting the role of dissociation in mediating the relationship between trauma and voices and evidence supporting the role of adult attachment in influencing beliefs and relationships that voice hearers can develop with voices. The paper concludes by outlining the key questions that future research needs to address to fully test the model and the clinical implications that arise from the work. PMID:28713292

  6. Does the vestibular system contribute to head direction cell activity in the rat?

    NASA Technical Reports Server (NTRS)

    Brown, J. E.; Yates, B. J.; Taube, J. S.; Oman, C. M. (Principal Investigator)

    2002-01-01

    Head direction cells (HDC) located in several regions of the brain, including the anterior dorsal nucleus of the thalamus (ADN), postsubiculum (PoS), and lateral mammillary nuclei (LMN), provide the neural substrate for the determination of head direction. Although activity of HDC is influenced by various sensory signals and internally generated cues, lesion studies and some anatomical and physiological evidence suggest that vestibular inputs are critical for the maintenance of directional sensitivity of these cells. However, vestibular inputs must be transformed considerably in order to signal head direction, and the neuronal circuitry that accomplishes this signal processing has not been fully established. Furthermore, it is unclear why the removal of vestibular inputs abolishes the directional sensitivity of HDC, as visual and other sensory inputs and motor feedback signals strongly affect the firing of these neurons and would be expected to maintain their directional-related activity. Further physiological studies will be required to establish the role of vestibular system in producing HDC responses, and anatomical studies are needed to determine the neural circuitry that mediates vestibular influences on determination of head direction.

  7. A Comparative Study of the VHI-10 and the V-RQOL for Quality of Life Among Chinese Teachers With and Without Voice Disorders.

    PubMed

    Lu, Dan; Wen, Bei; Yang, Hui; Chen, Fei; Liu, Jun; Xu, Yanan; Zheng, Yitao; Zhao, Yu; Zou, Jian; Wang, Haiyang

    2017-07-01

    To investigate the differences and correlation between the Voice Handicap Index-10 (VHI-10) and the Voice-Related Quality of Life (V-RQOL) in teachers in China with and without voice disorders. This is a cross-sectional descriptive analytical study. The participants were 864 teachers (569 women, 295 men) whose vocal cords were examined using a flexible nasofibrolaryngoscope. Questionnaire results were obtained for both the VHI-10 and the V-RQOL. Of the 864 participants, 409 teachers had no voice disorders and 455 teachers had voice disorders. The most common voice complaint was hoarseness (n = 298) and the most common throat complaint was globus pharyngis (n = 79) in teachers with voice disorders. Chronic laryngitis (n = 218) and polyps and nodules (n = 182) were the most frequent diagnoses in teachers with voice disorders. Significant differences were seen on the VHI-10 between teachers with and those without voice disorders (P < 0.05) and in function between female and male teachers with voice disorders (P < 0.05) and between those with different voice disorders (P < 0.05). Moderate to strong correlations were observed between VHI-10 total score and those for the three domains of the VHI-10 and the V-RQOL (P < 0.0001). There is a high prevalence of voice disorders in teachers. Teachers with voice disorders have poor voice-related quality of life, with more impairment seen among female than male teachers. Different groups of voice disorders have different effects on voice-related quality of life. A moderate correlation was found between the results of the VHI-10 and the V-RQOL. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  8. Voice hearing within the context of hearers' social worlds: an interpretative phenomenological analysis.

    PubMed

    Mawson, Amy; Berry, Katherine; Murray, Craig; Hayward, Mark

    2011-09-01

    Research has found relational qualities of power and intimacy to exist within hearer-voice interactions. The present study aimed to provide a deeper understanding of the interpersonal context of voice hearing by exploring participants' relationships with their voices and other people in their lives. This research was designed in consultation with service users and employed a qualitative, phenomenological, and idiographic design using semi-structured interviews. Ten participants, recruited via mental health services, and who reported hearing voices in the previous week, completed the interviews. These were transcribed verbatim and analysed using interpretative phenomenological analysis. Five themes resulted from the analysis. Theme 1: 'person and voice' demonstrated that participants' voices often reflected the identity, but not always the quality of social acquaintances. Theme 2: 'voices changing and confirming relationship with the self' explored the impact of voice hearing in producing an inferior sense-of-self in comparison to others. Theme 3: 'a battle for control' centred on issues of control and a dilemma of independence within voice relationships. Theme 4: 'friendships facilitating the ability to cope' and theme 5: 'voices creating distance in social relationships' explored experiences of social relationships within the context of voice hearing, and highlighted the impact of social isolation for voice hearers. The study demonstrated the potential role of qualitative research in developing theories of voice hearing. It extended previous research by highlighting the interface between voices and the social world of the hearer, including reciprocal influences of social relationships on voices and coping. Improving voice hearers' sense-of-self may be a key factor in reducing the distress caused by voices. ©2010 The British Psychological Society.

  9. Voice Disorders in Teacher Students-A Prospective Study and a Randomized Controlled Trial.

    PubMed

    Ohlsson, Ann-Christine; Andersson, Eva M; Södersten, Maria; Simberg, Susanna; Claesson, Silwa; Barregård, Lars

    2016-11-01

    Teachers are at risk of developing voice disorders, but longitudinal studies on voice problems among teachers are lacking. The aim of this randomized trial was to investigate long-term effects of voice education for teacher students with mild voice problems. In addition, vocal health was examined prospectively in a group of students without voice problems. First-semester students answered three questionnaires: one about background factors, one about voice symptoms (Screen6), and the Voice Handicap Index. Students with voice problems according to the questionnaire results were randomized to a voice training group or a control group. At follow-up in the sixth semester, all students answered Screen6 again together with four questions about factors that could have affected vocal health during their teacher education. The training group and the control group also answered the Voice Handicap Index a second time. At follow-up, 400 students remained in the study: 27 in the training group, 54 in the control group, and 319 without voice problems at baseline. Voice problems had decreased somewhat more in the training group than in the control group, but the difference was not statistically significant (P = 0.1). However, subgroup analyses showed significantly larger improvement among the students in the group with complete participation in the training program compared with the group with incomplete participation. Of the 319 students without voice problems at baseline, 14% had developed voice problems. Voice problems often develop in teacher students. Despite extensive dropout, our results support the hypothesis that voice education for teacher students has a preventive effect. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  10. Alerting prefixes for speech warning messages. [in helicopters

    NASA Technical Reports Server (NTRS)

    Bucher, N. M.; Voorhees, J. W.; Karl, R. L.; Werner, E.

    1984-01-01

    A major question posed by the design of an integrated voice information display/warning system for next-generation helicopter cockpits is whether an alerting prefix should precede voice warning messages; if so, the characteristics desirable in such a cue must also be addressed. Attention is presently given to the results of a study which ascertained pilot response time and response accuracy to messages preceded by either neutral cues or the cognitively appropriate semantic cues. Both verbal cues and messages were spoken in direct, phoneme-synthesized speech, and a training manipulation was included to determine the extent to which previous exposure to speech thus produced facilitates these messages' comprehension. Results are discussed in terms of the importance of human factors research in cockpit display design.

  11. Procedural Fairness and Creativity: Does Voice Maintain People's Creative Vein over Time?

    ERIC Educational Resources Information Center

    Streicher, Bernhard; Jonas, Eva; Maier, Günter W.; Frey, Dieter; Spießberger, Anneliese

    2012-01-01

    Although some research suggests a link between procedural fairness and creativity, so far no study has directly tested whether a real manipulation of procedural fairness affects creativity. Additionally, research on procedural fairness effects consists mostly of unique studies, but more realistic, life-like longitudinal experiments with repeated…

  12. Voices from the Community: A Case for Reciprocity in Service-Learning

    ERIC Educational Resources Information Center

    d'Arlach, Lucia; Sanchez, Bernadette; Feuer, Rachel

    2009-01-01

    Few studies have directly examined how recipients of service view the service. This qualitative study presents the results of interviews and observations of nine community members who participated in a service-learning, language exchange program, Intercambio, in which Spanish-speaking Latino immigrants were paired with English-speaking university…

  13. New Visions, New Voices: Future Directions in the Care and Education of Young Children.

    ERIC Educational Resources Information Center

    Williams, Leslie R.

    1989-01-01

    This article outlines key issues in early childhood education related to (1) identification and characterization of the populations to be served, (2) definition of the goals of services, (3) preparation of early childhood specialists, and (4) optimal settings for delivery of service. (IAH)

  14. Muscle Bioenergetic Considerations for Intrinsic Laryngeal Skeletal Muscle Physiology

    ERIC Educational Resources Information Center

    Sandage, Mary J.; Smith, Audrey G.

    2017-01-01

    Purpose: Intrinsic laryngeal skeletal muscle bioenergetics, the means by which muscles produce fuel for muscle metabolism, is an understudied aspect of laryngeal physiology with direct implications for voice habilitation and rehabilitation. The purpose of this review is to describe bioenergetic pathways identified in limb skeletal muscle and…

  15. Addressing Cyberconduct: A Brief to the Department of Justice Canada

    ERIC Educational Resources Information Center

    Canadian Teachers' Federation (NJ1), 2008

    2008-01-01

    This paper focuses on issues related to the misuse and abuse of the technologies, or "cybermisconduct" directed toward students and teachers such as online harassment, cyberbullying and internet defamation. Since some forms of cyberbullying may be criminal acts under the Criminal Code, Canadian courts are voicing serious concerns about…

  16. The Student Voice in Quality Assurance

    ERIC Educational Resources Information Center

    McKinney, Margaret; Comadina Granson, Ruben

    2012-01-01

    As lecturers of English and Spanish for international students in higher education, we frequently read reflections supporting the notion that the interaction between teacher and student has a direct influence on their learning and possibly even on their future studies and lives. Using examples taken from language learning histories, student course…

  17. Indian Voices; The First Convocation of American Indian Scholars.

    ERIC Educational Resources Information Center

    Costo, Rupert; And Others

    The document reports on The First Convocation of American Indian Scholars, which was attended by professional people, artists, traditional historians, etc. As noted, the 4-day convocation was conceived, organized, and directed entirely by Native Americans and was limited to 200 participants, among whom were 36 Native American students. The…

  18. Sound Solutions

    ERIC Educational Resources Information Center

    Starkman, Neal

    2007-01-01

    Poor classroom acoustics are impairing students' hearing and their ability to learn. However, technology has come up with a solution: tools that focus voices in a way that minimizes intrusive ambient noise and gets to the intended receiver--not merely amplifying the sound, but also clarifying and directing it. One provider of classroom audio…

  19. Articulatory Changes Following Treatment of Muscle Tension Dysphonia: Preliminary Acoustic Evidence

    ERIC Educational Resources Information Center

    Dromey, Christopher; Nissen, Shawn L.; Roy, Nelson; Merrill, Ray M.

    2008-01-01

    Purpose: Primary muscle tension dysphonia (MTD), a voice disturbance that occurs in the absence of structural or neurological pathology, may respond to manual circumlaryngeal techniques, which ostensibly alter the posture of the larynx and/or the configuration of the vocal folds without directly targeting supralaryngeal articulatory structures.…

  20. Defense.gov Special Report: Traumatic Brain Injury

    Science.gov Websites

    Excellence TBI Resources Brainline Military The Michael E. DeBakey VA Medical Center Congressionally Directed Medical Research Program NIH: National Institute of Neurological Disorders NIH: Traumatic Brain Injury Research CDC: Give Brain Injury a Voice Center for Medical Excellence for Multimedia Brainline.org - Brain

  1. 75 FR 43446 - Nondiscrimination on the Basis of Disability in State and Local Government Services...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-07-26

    ... text, pictures, and video) capabilities. This ANPRM seeks information on possible revisions to the... as text, pictures, and video) capabilities so that they will be able to directly receive various kinds of voice-, text- and video-based ``calls.'' Several commenters, including the National Emergency...

  2. Black Education: A Transformative Research and Action Agenda for the New Century

    ERIC Educational Resources Information Center

    King, Joyce E., Ed.

    2005-01-01

    This volume presents the findings and recommendations of the American Educational Research Association's (AERA) Commission on Research in Black Education (CORIBE) and offers new directions for research and practice. By commissioning an independent group of scholars of diverse perspectives and voices to investigate major issues hindering the…

  3. Mirror, Mirror on the Wall...?

    ERIC Educational Resources Information Center

    Pflaster, Gail

    1979-01-01

    The study determined the value of using a mirror for speech teaching by recording manner, place, voicing, and blend errors produced by 27 hearing-impaired children (5-13 years old) while imitating consonant-vowel syllables under three conditions (audition alone, audition plus direct vision, and audition plus vision using a mirror). (Author)

  4. Digital, Satellite-Based Aeronautical Communication

    NASA Technical Reports Server (NTRS)

    Davarian, F.

    1989-01-01

    Satellite system relays communication between aircraft and stations on ground. System offers better coverage with direct communication between air and ground, costs less and makes possible new communication services. Carries both voice and data. Because many data exchanged between aircraft and ground contain safety-related information, probability of bit errors essential.

  5. Direct connections assist neurons to detect correlation in small amplitude noises

    PubMed Central

    Bolhasani, E.; Azizi, Y.; Valizadeh, A.

    2013-01-01

    We address a question on the effect of common stochastic inputs on the correlation of the spike trains of two neurons when they are coupled through direct connections. We show that the change in the correlation of small amplitude stochastic inputs can be better detected when the neurons are connected by direct excitatory couplings. Depending on whether intrinsic firing rate of the neurons is identical or slightly different, symmetric or asymmetric connections can increase the sensitivity of the system to the input correlation by changing the mean slope of the correlation transfer function over a given range of input correlation. In either case, there is also an optimum value for synaptic strength which maximizes the sensitivity of the system to the changes in input correlation. PMID:23966940

  6. Quantitative evaluation of the voice range profile in patients with voice disorder.

    PubMed

    Ikeda, Y; Masuda, T; Manako, H; Yamashita, H; Yamamoto, T; Komiyama, S

    1999-01-01

    In 1953, Calvet first displayed the fundamental frequency (pitch) and sound pressure level (intensity) of a voice on a two-dimensional plane and created a voice range profile. This profile has been used to evaluate clinically various vocal disorders, although such evaluations to date have been subjective without quantitative assessment. In the present study, a quantitative system was developed to evaluate the voice range profile utilizing a personal computer. The area of the voice range profile was defined as the voice volume. This volume was analyzed in 137 males and 175 females who were treated for various dysphonias at Kyushu University between 1984 and 1990. Ten normal subjects served as controls. The voice volume in cases with voice disorders significantly decreased irrespective of the disease and sex. Furthermore, cases having better improvement after treatment showed a tendency for the voice volume to increase. These findings illustrated the voice volume as a useful clinical test for evaluating voice control in cases with vocal disorders.

  7. Unmasking the effects of masking on performance: The potential of multiple-voice masking in the office environment.

    PubMed

    Keus van de Poll, Marijke; Carlsson, Johannes; Marsh, John E; Ljung, Robert; Odelius, Johan; Schlittmeier, Sabine J; Sundin, Gunilla; Sörqvist, Patrik

    2015-08-01

    Broadband noise is often used as a masking sound to combat the negative consequences of background speech on performance in open-plan offices. As office workers generally dislike broadband noise, it is important to find alternatives that are more appreciated while being at least not less effective. The purpose of experiment 1 was to compare broadband noise with two alternatives-multiple voices and water waves-in the context of a serial short-term memory task. A single voice impaired memory in comparison with silence, but when the single voice was masked with multiple voices, performance was on level with silence. Experiment 2 explored the benefits of multiple-voice masking in more detail (by comparing one voice, three voices, five voices, and seven voices) in the context of word processed writing (arguably a more office-relevant task). Performance (i.e., writing fluency) increased linearly from worst performance in the one-voice condition to best performance in the seven-voice condition. Psychological mechanisms underpinning these effects are discussed.

  8. High-speed asynchronous data mulitiplexer/demultiplexer for high-density digital recorders

    NASA Astrophysics Data System (ADS)

    Berdugo, Albert; Small, Martin B.

    1996-11-01

    Modern High Density Digital Recorders are ideal devices for the storage of large amounts of digital and/or wideband analog data. Ruggedized versions of these recorders are currently available and are supporting many military and commercial flight test applications. However, in certain cases, the storage format becomes very critical, e.g., when a large number of data types are involved, or when channel- to-channel correlation is critical, or when the original data source must be accurately recreated during post mission analysis. A properly designed storage format will not only preserve data quality, but will yield the maximum storage capacity and record time for any given recorder family or data type. This paper describes a multiplex/demultiplex technique that formats multiple high speed data sources into a single, common format for recording. The method is compatible with many popular commercial recorder standards such as DCRsi, VLDS, and DLT. Types of input data typically include PCM, wideband analog data, video, aircraft data buses, avionics, voice, time code, and many others. The described method preserves tight data correlation with minimal data overhead. The described technique supports full reconstruction of the original input signals during data playback. Output data correlation across channels is preserved for all types of data inputs. Simultaneous real- time data recording and reconstruction are also supported.

  9. The Effects of Perceptions of Organizational Structure on Job Involvement, Job Satisfaction, and Organizational Commitment Among Indian Police Officers.

    PubMed

    Lambert, Eric G; Qureshi, Hanif; Klahm, Charles; Smith, Brad; Frank, James

    2017-12-01

    Successful police organizations rely on involved, satisfied, and committed workers. The concepts of job involvement (i.e., connection with the job), job satisfaction (i.e., affective feeling toward the job), and organizational commitment (i.e., bond with the employing organization) have been shown to significantly affect intentions and behaviors of employees. The current study used multivariate ordinary least squares (OLS) regression analysis on survey results from a sample of 827 Indian police officers to explore how perceptions of work environment factors affect officers' job involvement, job satisfaction, and organizational commitment. Organizational support, formalization (i.e., level of codified written rules and guidelines), promotional opportunities, institutional communication (i.e., salient work information is transmitted), and input into decision-making (i.e., having a voice in the process) significantly influenced the job involvement, job satisfaction, and organizational commitment of Indian police officers. Specifically, in the multivariate analysis, perceptions of formalization and instrumental communication had a positive relationship with job involvement; perceptions of organizational support, promotional opportunities, instrumental communication, and input into decision-making had positive associations with job satisfaction; and perceptions of organizational support, formalization, promotional opportunities, instrumental communication, and input into decision-making had positive relationships with organizational commitment.

  10. Power and resistance within the hospital's hierarchical system: the experiences of chronically ill patients.

    PubMed

    Griscti, Odette; Aston, Megan; Warner, Grace; Martin-Misener, Ruth; McLeod, Deborah

    2017-01-01

    To explore experiences of chronically ill patients and registered nurses when they negotiate patient care in hospital settings. Specifically, we explored how social and institutional discourses shape power relations during the negotiation process. The hospital system is embedded in a hierarchical structure where the voice of the healthcare provider as expert is often given more importance than the patient. This system has been criticised as being oppressive to patients who are perceived to be lower in the hierarchy. In this study, we illustrate how the hospital's hierarchical system is not always oppressing but can also create moments of empowerment for patients. A feminist poststructuralist approach informed by the teaching of Foucault was used to explore power relations between nurses and patients when negotiating patient care in hospital settings. Eight individuals who suffered from chronic illness shared their stories about how they negotiated their care with nurses in hospital settings. The interviews were tape-recorded. Discourse analysis was used to analyse the data. Patients recounted various experiences when their voices were not heard because the current hospital system privileged the healthcare provider experts' advice over the patients' voice. The hierarchical structure of hospital supported these dynamics by privileging nurses as gatekeepers of service, by excluding the patients' input in the nursing notes and through a process of self-regulation. However, patients in this study were not passive recipients of care and used their agency creatively to resist these discourses. Nurses need to be mindful of how the hospital's hierarchical system tends to place nurses in a position of power, and how their authoritative position may positively or adversely affect the negotiation of patient care. © 2016 John Wiley & Sons Ltd.

  11. Constraints on the Transfer of Perceptual Learning in Accented Speech

    PubMed Central

    Eisner, Frank; Melinger, Alissa; Weber, Andrea

    2013-01-01

    The perception of speech sounds can be re-tuned through a mechanism of lexically driven perceptual learning after exposure to instances of atypical speech production. This study asked whether this re-tuning is sensitive to the position of the atypical sound within the word. We investigated perceptual learning using English voiced stop consonants, which are commonly devoiced in word-final position by Dutch learners of English. After exposure to a Dutch learner’s productions of devoiced stops in word-final position (but not in any other positions), British English (BE) listeners showed evidence of perceptual learning in a subsequent cross-modal priming task, where auditory primes with devoiced final stops (e.g., “seed”, pronounced [si:th]), facilitated recognition of visual targets with voiced final stops (e.g., SEED). In Experiment 1, this learning effect generalized to test pairs where the critical contrast was in word-initial position, e.g., auditory primes such as “town” facilitated recognition of visual targets like DOWN. Control listeners, who had not heard any stops by the speaker during exposure, showed no learning effects. The generalization to word-initial position did not occur when participants had also heard correctly voiced, word-initial stops during exposure (Experiment 2), and when the speaker was a native BE speaker who mimicked the word-final devoicing (Experiment 3). The readiness of the perceptual system to generalize a previously learned adjustment to other positions within the word thus appears to be modulated by distributional properties of the speech input, as well as by the perceived sociophonetic characteristics of the speaker. The results suggest that the transfer of pre-lexical perceptual adjustments that occur through lexically driven learning can be affected by a combination of acoustic, phonological, and sociophonetic factors. PMID:23554598

  12. Speed control system for an access gate

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bzorgi, Fariborz M

    2012-03-20

    An access control apparatus for an access gate. The access gate typically has a rotator that is configured to rotate around a rotator axis at a first variable speed in a forward direction. The access control apparatus may include a transmission that typically has an input element that is operatively connected to the rotator. The input element is generally configured to rotate at an input speed that is proportional to the first variable speed. The transmission typically also has an output element that has an output speed that is higher than the input speed. The input element and the outputmore » element may rotate around a common transmission axis. A retardation mechanism may be employed. The retardation mechanism is typically configured to rotate around a retardation mechanism axis. Generally the retardation mechanism is operatively connected to the output element of the transmission and is configured to retard motion of the access gate in the forward direction when the first variable speed is above a control-limit speed. In many embodiments the transmission axis and the retardation mechanism axis are substantially co-axial. Some embodiments include a freewheel/catch mechanism that has an input connection that is operatively connected to the rotator. The input connection may be configured to engage an output connection when the rotator is rotated at the first variable speed in a forward direction and configured for substantially unrestricted rotation when the rotator is rotated in a reverse direction opposite the forward direction. The input element of the transmission is typically operatively connected to the output connection of the freewheel/catch mechanism.« less

  13. Voice Tremor in Parkinson's Disease: An Acoustic Study.

    PubMed

    Gillivan-Murphy, Patricia; Miller, Nick; Carding, Paul

    2018-01-30

    Voice tremor associated with Parkinson disease (PD) has not been characterized. Its relationship with voice disability and disease variables is unknown. This study aimed to evaluate voice tremor in people with PD (pwPD) and a matched control group using acoustic analysis, and to examine correlations with voice disability and disease variables. Acoustic voice tremor analysis was completed on 30 pwPD and 28 age-gender matched controls. Voice disability (Voice Handicap Index), and disease variables of disease duration, Activities of Daily Living (Unified Parkinson's Disease Rating Scale [UPDRS II]), and motor symptoms related to PD (UPDRS III) were examined for relationship with voice tremor measures. Voice tremor was detected acoustically in pwPD and controls with similar frequency. PwPD had a statistically significantly higher rate of amplitude tremor (Hz) than controls (P = 0.001). Rate of amplitude tremor was negatively and significantly correlated with UPDRS III total score (rho -0.509). For pwPD, the magnitude and periodicity of acoustic tremor was higher than for controls without statistical significance. The magnitude of frequency tremor (Mftr%) was positively and significantly correlated with disease duration (rho 0.463). PwPD had higher Voice Handicap Index total, functional, emotional, and physical subscale scores than matched controls (P < 0.001). Voice disability did not correlate significantly with acoustic voice tremor measures. Acoustic analysis enhances understanding of PD voice tremor characteristics, its pathophysiology, and its relationship with voice disability and disease symptomatology. Copyright © 2018 The Voice Foundation. All rights reserved.

  14. Epidemiology of Voice Disorders in Latvian School Teachers.

    PubMed

    Trinite, Baiba

    2017-07-01

    The prevalence of voice disorders in the teacher population in Latvia has not been studied so far and this is the first epidemiological study whose goal is to investigate the prevalence of voice disorders and their risk factors in this professional group. A wide cross-sectional study using stratified sampling methodology was implemented in the general education schools of Latvia. The self-administered voice risk factor questionnaire and the Voice Handicap Index were completed by 522 teachers. Two teachers groups were formed: the voice disorders group which included 235 teachers with actual voice problems or problems during the last 9 months; and the control group which included 174 teachers without voice disorders. Sixty-six percent of teachers gave a positive answer to the following question: Have you ever had problems with your voice? Voice problems are more often found in female than male teachers (68.2% vs 48.8%). Music teachers suffer from voice disorders more often than teachers of other subjects. Eighty-two percent of teachers first faced voice problems in their professional carrier. The odds of voice disorders increase if the following risk factors exist: extra vocal load, shouting, throat clearing, neglecting of personal health, background noise, chronic illnesses of the upper respiratory tract, allergy, job dissatisfaction, and regular stress in the working place. The study findings indicated a high risk of voice disorders among Latvian teachers. The study confirmed data concerning the multifactorial etiology of voice disorders. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  15. Speaker's comfort in teaching environments: voice problems in Swedish teaching staff.

    PubMed

    Åhlander, Viveka Lyberg; Rydell, Roland; Löfqvist, Anders

    2011-07-01

    The primary objective of this study was to examine how a group of Swedish teachers rate aspects of their working environment that can be presumed to have an impact on vocal behavior and voice problems. The secondary objective was to explore the prevalence of voice problems in Swedish teachers. Questionnaires were distributed to the teachers of 23 randomized schools. Teaching staff at all levels were included, except preschool teachers and teachers at specialized, vocational high schools. The response rate was 73%. The results showed that 13% of the whole group reported voice problems occurring sometimes, often, or always. The teachers reporting voice problems were compared with those without problems. There were significant differences among the groups for several items. The teachers with voice problems rated items on room acoustics and work environment as more noticeable. This group also reported voice symptoms, such as hoarseness, throat clearing, and voice change, to a significantly higher degree, even though teachers in both groups reported some voice symptoms. Absence from work because of voice problems was also significantly more common in the group with voice problems--35% versus 9% in the group without problems. We may conclude that teachers suffering from voice problems react stronger to loading factors in the teaching environment, report more frequent symptoms of voice discomfort, and are more often absent from work because of voice problems than their voice-healthy colleagues. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  16. Selective attention in perceptual adjustments to voice.

    PubMed

    Mullennix, J W; Howe, J N

    1999-10-01

    The effects of perceptual adjustments to voice information on the perception of isolated spoken words were examined. In two experiments, spoken target words were preceded or followed within a trial by a neutral word spoken in the same voice or in a different voice as the target. Over-all, words were reproduced more accurately on trials on which the voice of the neutral word matched the voice of the spoken target word, suggesting that perceptual adjustments to voice interfere with word processing. This result, however, was mediated by selective attention to voice. The results provide further evidence of a close processing relationship between perceptual adjustments to voice and spoken word recognition.

  17. Passively damped vibration welding system and method

    DOEpatents

    Tan, Chin-An; Kang, Bongsu; Cai, Wayne W.; Wu, Tao

    2013-04-02

    A vibration welding system includes a controller, welding horn, an anvil, and a passive damping mechanism (PDM). The controller generates an input signal having a calibrated frequency. The horn vibrates in a desirable first direction at the calibrated frequency in response to the input signal to form a weld in a work piece. The PDM is positioned with respect to the system, and substantially damps or attenuates vibration in an undesirable second direction. A method includes connecting the PDM having calibrated properties and a natural frequency to an anvil of an ultrasonic welding system. Then, an input signal is generated using a weld controller. The method includes vibrating a welding horn in a desirable direction in response to the input signal, and passively damping vibration in an undesirable direction using the PDM.

  18. Understanding the 'Anorexic Voice' in Anorexia Nervosa.

    PubMed

    Pugh, Matthew; Waller, Glenn

    2017-05-01

    In common with individuals experiencing a number of disorders, people with anorexia nervosa report experiencing an internal 'voice'. The anorexic voice comments on the individual's eating, weight and shape and instructs the individual to restrict or compensate. However, the core characteristics of the anorexic voice are not known. This study aimed to develop a parsimonious model of the voice characteristics that are related to key features of eating disorder pathology and to determine whether patients with anorexia nervosa fall into groups with different voice experiences. The participants were 49 women with full diagnoses of anorexia nervosa. Each completed validated measures of the power and nature of their voice experience and of their responses to the voice. Different voice characteristics were associated with current body mass index, duration of disorder and eating cognitions. Two subgroups emerged, with 'weaker' and 'stronger' voice experiences. Those with stronger voices were characterized by having more negative eating attitudes, more severe compensatory behaviours, a longer duration of illness and a greater likelihood of having the binge-purge subtype of anorexia nervosa. The findings indicate that the anorexic voice is an important element of the psychopathology of anorexia nervosa. Addressing the anorexic voice might be helpful in enhancing outcomes of treatments for anorexia nervosa, but that conclusion might apply only to patients with more severe eating psychopathology. Copyright © 2016 John Wiley & Sons, Ltd. Experiences of an internal 'anorexic voice' are common in anorexia nervosa. Clinicians should consider the role of the voice when formulating eating pathology in anorexia nervosa, including how individuals perceive and relate to that voice. Addressing the voice may be beneficial, particularly in more severe and enduring forms of anorexia nervosa. When working with the voice, clinicians should aim to address both the content of the voice and how individuals relate and respond to it. Copyright © 2016 John Wiley & Sons, Ltd.

  19. 14 CFR 23.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 14 Aeronautics and Space 1 2011-01-01 2011-01-01 false Cockpit voice recorders. 23.1457 Section 23... Equipment § 23.1457 Cockpit voice recorders. (a) Each cockpit voice recorder required by the operating rules...) Voice communications transmitted from or received in the airplane by radio. (2) Voice communications of...

  20. Voices Not Heard: Voice-Use Profiles of Elementary Music Teachers, the Effects of Voice Amplification on Vocal Load, and Perceptions of Issues Surrounding Voice Use

    ERIC Educational Resources Information Center

    Morrow, Sharon L.

    2009-01-01

    Teachers represent the largest group of occupational voice users and have voice-related problems at a rate of over twice that found in the general population. Among teachers, music teachers are roughly four times more likely than classroom teachers to develop voice-related problems. Although it has been established that music teachers use their…

  1. Cognitive Behavioural Relating Therapy (CBRT) for voice hearers: a case study.

    PubMed

    Paulik, Georgie; Hayward, Mark; Birchwood, Max

    2013-10-01

    There has been a recent focus on the interpersonal nature of the voice hearing experience, with studies showing that similar patterns of relating exist between voice hearer and voice as between voice hearer and social others. Two recent therapeutic approaches to voices, Cognitive Therapy for Command Hallucinations and Relating Therapy, have been developed to address patterns of relating and power imbalances between voice hearer and voice. This paper presents a novel intervention that combines elements of these two therapies, named Cognitive Behavioural Relating Therapy (CBRT). The application of CBRT is illustrated through a clinical case study. The clinical case study showed changes in patterns of relating, improved self-esteem and reductions in voice-related distress. The outcomes provide preliminary support for the utility of CBRT when working with voice hearers.

  2. Voice characteristics of children aged between 6 and 13 years: impact of age, gender, and vocal training.

    PubMed

    Pribuisiene, Ruta; Uloza, Virgilijus; Kardisiene, Vilija

    2011-12-01

    To determine impact of age, gender, and vocal training on voice characteristics of children aged 6-13 years. Voice acoustic and phonetogram parameters were determined for the group of 44 singing and 31 non-singing children. No impact of gender and/or age on phonetogram, acoustic voice parameters, and maximum phonation time was detected. Voice ranges of all children represented a pre-pubertal soprano type with a voice range of 22 semitones for non-singing and of 26 semitones for singing individuals. The mean maximum voice intensity was 81 dB. Vocal training had a positive impact on voice intensity parameters in girls. The presented data on average voice characteristics may be applicable in the clinical practice and provide relevant support for voice assessment.

  3. Understanding the mechanisms of familiar voice-identity recognition in the human brain.

    PubMed

    Maguinness, Corrina; Roswandowitz, Claudia; von Kriegstein, Katharina

    2018-03-31

    Humans have a remarkable skill for voice-identity recognition: most of us can remember many voices that surround us as 'unique'. In this review, we explore the computational and neural mechanisms which may support our ability to represent and recognise a unique voice-identity. We examine the functional architecture of voice-sensitive regions in the superior temporal gyrus/sulcus, and bring together findings on how these regions may interact with each other, and additional face-sensitive regions, to support voice-identity processing. We also contrast findings from studies on neurotypicals and clinical populations which have examined the processing of familiar and unfamiliar voices. Taken together, the findings suggest that representations of familiar and unfamiliar voices might dissociate in the human brain. Such an observation does not fit well with current models for voice-identity processing, which by-and-large assume a common sequential analysis of the incoming voice signal, regardless of voice familiarity. We provide a revised audio-visual integrative model of voice-identity processing which brings together traditional and prototype models of identity processing. This revised model includes a mechanism of how voice-identity representations are established and provides a novel framework for understanding and examining the potential differences in familiar and unfamiliar voice processing in the human brain. Copyright © 2018 Elsevier Ltd. All rights reserved.

  4. Vocal Age Disguise: The Role of Fundamental Frequency and Speech Rate and Its Perceived Effects

    PubMed Central

    Skoog Waller, Sara; Eriksson, Mårten

    2016-01-01

    The relationship between vocal characteristics and perceived age is of interest in various contexts, as is the possibility to affect age perception through vocal manipulation. A few examples of such situations are when age is staged by actors, when ear witnesses make age assessments based on vocal cues only or when offenders (e.g., online groomers) disguise their voice to appear younger or older. This paper investigates how speakers spontaneously manipulate two age related vocal characteristics (f0 and speech rate) in attempt to sound younger versus older than their true age, and if the manipulations correspond to actual age related changes in f0 and speech rate (Study 1). Further aims of the paper is to determine how successful vocal age disguise is by asking listeners to estimate the age of generated speech samples (Study 2) and to examine whether or not listeners use f0 and speech rate as cues to perceived age. In Study 1, participants from three age groups (20–25, 40–45, and 60–65 years) agreed to read a short text under three voice conditions. There were 12 speakers in each age group (six women and six men). They used their natural voice in one condition, attempted to sound 20 years younger in another and 20 years older in a third condition. In Study 2, 60 participants (listeners) listened to speech samples from the three voice conditions in Study 1 and estimated the speakers’ age. Each listener was exposed to all three voice conditions. The results from Study 1 indicated that the speakers increased fundamental frequency (f0) and speech rate when attempting to sound younger and decreased f0 and speech rate when attempting to sound older. Study 2 showed that the voice manipulations had an effect in the sought-after direction, although the achieved mean effect was only 3 years, which is far less than the intended effect of 20 years. Moreover, listeners used speech rate, but not f0, as a cue to speaker age. It was concluded that age disguise by voice can be achieved by naïve speakers even though the perceived effect was smaller than intended. PMID:27917144

  5. Voices to reckon with: perceptions of voice identity in clinical and non-clinical voice hearers

    PubMed Central

    Badcock, Johanna C.; Chhabra, Saruchi

    2013-01-01

    The current review focuses on the perception of voice identity in clinical and non-clinical voice hearers. Identity perception in auditory verbal hallucinations (AVH) is grounded in the mechanisms of human (i.e., real, external) voice perception, and shapes the emotional (distress) and behavioral (help-seeking) response to the experience. Yet, the phenomenological assessment of voice identity is often limited, for example to the gender of the voice, and has failed to take advantage of recent models and evidence on human voice perception. In this paper we aim to synthesize the literature on identity in real and hallucinated voices and begin by providing a comprehensive overview of the features used to judge voice identity in healthy individuals and in people with schizophrenia. The findings suggest some subtle, but possibly systematic biases across different levels of voice identity in clinical hallucinators that are associated with higher levels of distress. Next we provide a critical evaluation of voice processing abilities in clinical and non-clinical voice hearers, including recent data collected in our laboratory. Our studies used diverse methods, assessing recognition and binding of words and voices in memory as well as multidimensional scaling of voice dissimilarity judgments. The findings overall point to significant difficulties recognizing familiar speakers and discriminating between unfamiliar speakers in people with schizophrenia, both with and without AVH. In contrast, these voice processing abilities appear to be generally intact in non-clinical hallucinators. The review highlights some important avenues for future research and treatment of AVH associated with a need for care, and suggests some novel insights into other symptoms of psychosis. PMID:23565088

  6. Exploring the feasibility of smart phone microphone for measurement of acoustic voice parameters and voice pathology screening.

    PubMed

    Uloza, Virgilijus; Padervinskis, Evaldas; Vegiene, Aurelija; Pribuisiene, Ruta; Saferis, Viktoras; Vaiciukynas, Evaldas; Gelzinis, Adas; Verikas, Antanas

    2015-11-01

    The objective of this study is to evaluate the reliability of acoustic voice parameters obtained using smart phone (SP) microphones and investigate the utility of use of SP voice recordings for voice screening. Voice samples of sustained vowel/a/obtained from 118 subjects (34 normal and 84 pathological voices) were recorded simultaneously through two microphones: oral AKG Perception 220 microphone and SP Samsung Galaxy Note3 microphone. Acoustic voice signal data were measured for fundamental frequency, jitter and shimmer, normalized noise energy (NNE), signal to noise ratio and harmonic to noise ratio using Dr. Speech software. Discriminant analysis-based Correct Classification Rate (CCR) and Random Forest Classifier (RFC) based Equal Error Rate (EER) were used to evaluate the feasibility of acoustic voice parameters classifying normal and pathological voice classes. Lithuanian version of Glottal Function Index (LT_GFI) questionnaire was utilized for self-assessment of the severity of voice disorder. The correlations of acoustic voice parameters obtained with two types of microphones were statistically significant and strong (r = 0.73-1.0) for the entire measurements. When classifying into normal/pathological voice classes, the Oral-NNE revealed the CCR of 73.7% and the pair of SP-NNE and SP-shimmer parameters revealed CCR of 79.5%. However, fusion of the results obtained from SP voice recordings and GFI data provided the CCR of 84.60% and RFC revealed the EER of 7.9%, respectively. In conclusion, measurements of acoustic voice parameters using SP microphone were shown to be reliable in clinical settings demonstrating high CCR and low EER when distinguishing normal and pathological voice classes, and validated the suitability of the SP microphone signal for the task of automatic voice analysis and screening.

  7. Learned face-voice pairings facilitate visual search.

    PubMed

    Zweig, L Jacob; Suzuki, Satoru; Grabowecky, Marcia

    2015-04-01

    Voices provide a rich source of information that is important for identifying individuals and for social interaction. During search for a face in a crowd, voices often accompany visual information, and they facilitate localization of the sought-after individual. However, it is unclear whether this facilitation occurs primarily because the voice cues the location of the face or because it also increases the salience of the associated face. Here we demonstrate that a voice that provides no location information nonetheless facilitates visual search for an associated face. We trained novel face-voice associations and verified learning using a two-alternative forced choice task in which participants had to correctly match a presented voice to the associated face. Following training, participants searched for a previously learned target face among other faces while hearing one of the following sounds (localized at the center of the display): a congruent learned voice, an incongruent but familiar voice, an unlearned and unfamiliar voice, or a time-reversed voice. Only the congruent learned voice speeded visual search for the associated face. This result suggests that voices facilitate the visual detection of associated faces, potentially by increasing their visual salience, and that the underlying crossmodal associations can be established through brief training.

  8. [Voice assessment and demographic data of applicants for a school of speech therapists].

    PubMed

    Reiter, R; Brosch, S

    2008-05-01

    Demographic data, subjective und objective voice analysis as well as self-assessment of voice quality from applicants for a school of speech therapists were investigated. Demographic data from 116 applicants were collected and their voice quality assessed by three independent judges. An objective evaluation was done by maximum phonation time, average fundamental frequency, dynamic range and percent of jitter and shimmer by means of Goettinger Hoarseness diagram. Self-assessment of voice quality was done by "voice handicap index questionnaire". The twenty successful applicants had a physiological voice in 95 %, they were all musical and had university entrance qualifications. Subjective voice assessment showed in 16 % of the applicants a hoarse voice. In this subgroup an unphysiological vocal use was observed in 72 % and a reduced articulation in 45 %. The objective voice parameters did not show a significant difference between the 3 groups. Self-assessment of the voice was inconspicuous in all applicants. Applicants with general qualification for university entrance, musicality and a physiological voice were more likely to be successful. There were main differences between self assessment of voice and quantitative analysis or subjective assessment by three independent judges.

  9. Keeping Your Voice Healthy

    MedlinePlus

    ... an ENT Doctor Near You Keeping Your Voice Healthy Keeping Your Voice Healthy Patient Health Information News ... voice-related. Key Steps for Keeping Your Voice Healthy Drink plenty of water. Moisture is good for ...

  10. Overgeneral autobiographical memory bias in clinical and non-clinical voice hearers.

    PubMed

    Jacobsen, Pamela; Peters, Emmanuelle; Ward, Thomas; Garety, Philippa A; Jackson, Mike; Chadwick, Paul

    2018-03-14

    Hearing voices can be a distressing and disabling experience for some, whilst it is a valued experience for others, so-called 'healthy voice-hearers'. Cognitive models of psychosis highlight the role of memory, appraisal and cognitive biases in determining emotional and behavioural responses to voices. A memory bias potentially associated with distressing voices is the overgeneral memory bias (OGM), namely the tendency to recall a summary of events rather than specific occasions. It may limit access to autobiographical information that could be helpful in re-appraising distressing experiences, including voices. We investigated the possible links between OGM and distressing voices in psychosis by comparing three groups: (1) clinical voice-hearers (N = 39), (2) non-clinical voice-hearers (N = 35) and (3) controls without voices (N = 77) on a standard version of the autobiographical memory test (AMT). Clinical and non-clinical voice-hearers also completed a newly adapted version of the task, designed to assess voices-related memories (vAMT). As hypothesised, the clinical group displayed an OGM bias by retrieving fewer specific autobiographical memories on the AMT compared with both the non-clinical and control groups, who did not differ from each other. The clinical group also showed an OGM bias in recall of voice-related memories on the vAMT, compared with the non-clinical group. Clinical voice-hearers display an OGM bias when compared with non-clinical voice-hearers on both general and voices-specific recall tasks. These findings have implications for the refinement and targeting of psychological interventions for psychosis.

  11. Randomized controlled trial of supplemental augmentative and alternative communication versus voice rest alone after phonomicrosurgery.

    PubMed

    Rousseau, Bernard; Gutmann, Michelle L; Mau, Theodore; Francis, David O; Johnson, Jeffrey P; Novaleski, Carolyn K; Vinson, Kimberly N; Garrett, C Gaelyn

    2015-03-01

    This randomized trial investigated voice rest and supplemental text-to-speech communication versus voice rest alone on visual analog scale measures of communication effectiveness and magnitude of voice use. Randomized clinical trial. Multicenter outpatient voice clinics. Thirty-seven patients undergoing phonomicrosurgery. Patients undergoing phonomicrosurgery were randomized to voice rest and supplemental text-to-speech communication or voice rest alone. The primary outcome measure was the impact of voice rest on ability to communicate effectively over a 7-day period. Pre- and postoperative magnitude of voice use was also measured as an observational outcome. Patients randomized to voice rest and supplemental text-to-speech communication reported higher median communication effectiveness on each postoperative day compared to those randomized to voice rest alone, with significantly higher median communication effectiveness on postoperative days 3 (P=.03) and 5 (P=.01). Magnitude of voice use did not differ on any preoperative (P>.05) or postoperative day (P>.05), nor did patients significantly decrease voice use as the surgery date approached (P>.05). However, there was a significant reduction in median voice use pre- to postoperatively across patients (P<.001) with median voice use ranging from 0 to 3 throughout the postoperative week. Supplemental text-to-speech communication increased patient-perceived communication effectiveness on postoperative days 3 and 5 over voice rest alone. With the prevalence of smartphones and the widespread use of text messaging, supplemental text-to-speech communication may provide an accessible and cost-effective communication option for patients on vocal restrictions. © American Academy of Otolaryngology—Head and Neck Surgery Foundation 2015.

  12. Peripheral hearing loss reduces the ability of children to direct selective attention during multi-talker listening.

    PubMed

    Holmes, Emma; Kitterick, Padraig T; Summerfield, A Quentin

    2017-07-01

    Restoring normal hearing requires knowledge of how peripheral and central auditory processes are affected by hearing loss. Previous research has focussed primarily on peripheral changes following sensorineural hearing loss, whereas consequences for central auditory processing have received less attention. We examined the ability of hearing-impaired children to direct auditory attention to a voice of interest (based on the talker's spatial location or gender) in the presence of a common form of background noise: the voices of competing talkers (i.e. during multi-talker, or "Cocktail Party" listening). We measured brain activity using electro-encephalography (EEG) when children prepared to direct attention to the spatial location or gender of an upcoming target talker who spoke in a mixture of three talkers. Compared to normally-hearing children, hearing-impaired children showed significantly less evidence of preparatory brain activity when required to direct spatial attention. This finding is consistent with the idea that hearing-impaired children have a reduced ability to prepare spatial attention for an upcoming talker. Moreover, preparatory brain activity was not restored when hearing-impaired children listened with their acoustic hearing aids. An implication of these findings is that steps to improve auditory attention alongside acoustic hearing aids may be required to improve the ability of hearing-impaired children to understand speech in the presence of competing talkers. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Children's Voice or Children's Voices? How Educational Research Can Be at the Heart of Schooling

    ERIC Educational Resources Information Center

    Stern, Julian

    2015-01-01

    There are problems with considering children and young people in schools as quite separate individuals, and with considering them as members of a single collectivity. The tension is represented in the use of "voice" and "voices" in educational debates. Voices in dialogue, in contrast to "children's voice", are…

  14. You're a What? Voice Actor

    ERIC Educational Resources Information Center

    Liming, Drew

    2009-01-01

    This article talks about voice actors and features Tony Oliver, a professional voice actor. Voice actors help to bring one's favorite cartoon and video game characters to life. They also do voice-overs for radio and television commercials and movie trailers. These actors use the sound of their voice to sell a character's emotions--or an advertised…

  15. Normal voice processing after posterior superior temporal sulcus lesion.

    PubMed

    Jiahui, Guo; Garrido, Lúcia; Liu, Ran R; Susilo, Tirta; Barton, Jason J S; Duchaine, Bradley

    2017-10-01

    The right posterior superior temporal sulcus (pSTS) shows a strong response to voices, but the cognitive processes generating this response are unclear. One possibility is that this activity reflects basic voice processing. However, several fMRI and magnetoencephalography findings suggest instead that pSTS serves as an integrative hub that combines voice and face information. Here we investigate whether right pSTS contributes to basic voice processing by testing Faith, a patient whose right pSTS was resected, with eight behavioral tasks assessing voice identity perception and recognition, voice sex perception, and voice expression perception. Faith performed normally on all the tasks. Her normal performance indicates right pSTS is not necessary for intact voice recognition and suggests that pSTS activations to voices reflect higher-level processes. Copyright © 2017 Elsevier Ltd. All rights reserved.

  16. A pneumatic Bionic Voice prosthesis-Pre-clinical trials of controlling the voice onset and offset.

    PubMed

    Ahmadi, Farzaneh; Noorian, Farzad; Novakovic, Daniel; van Schaik, André

    2018-01-01

    Despite emergent progress in many fields of bionics, a functional Bionic Voice prosthesis for laryngectomy patients (larynx amputees) has not yet been achieved, leading to a lifetime of vocal disability for these patients. This study introduces a novel framework of Pneumatic Bionic Voice Prostheses as an electronic adaptation of the Pneumatic Artificial Larynx (PAL) device. The PAL is a non-invasive mechanical voice source, driven exclusively by respiration with an exceptionally high voice quality, comparable to the existing gold standard of Tracheoesophageal (TE) voice prosthesis. Following PAL design closely as the reference, Pneumatic Bionic Voice Prostheses seem to have a strong potential to substitute the existing gold standard by generating a similar voice quality while remaining non-invasive and non-surgical. This paper designs the first Pneumatic Bionic Voice prosthesis and evaluates its onset and offset control against the PAL device through pre-clinical trials on one laryngectomy patient. The evaluation on a database of more than five hours of continuous/isolated speech recordings shows a close match between the onset/offset control of the Pneumatic Bionic Voice and the PAL with an accuracy of 98.45 ±0.54%. When implemented in real-time, the Pneumatic Bionic Voice prosthesis controller has an average onset/offset delay of 10 milliseconds compared to the PAL. Hence it addresses a major disadvantage of previous electronic voice prostheses, including myoelectric Bionic Voice, in meeting the short time-frames of controlling the onset/offset of the voice in continuous speech.

  17. A pneumatic Bionic Voice prosthesis—Pre-clinical trials of controlling the voice onset and offset

    PubMed Central

    Noorian, Farzad; Novakovic, Daniel; van Schaik, André

    2018-01-01

    Despite emergent progress in many fields of bionics, a functional Bionic Voice prosthesis for laryngectomy patients (larynx amputees) has not yet been achieved, leading to a lifetime of vocal disability for these patients. This study introduces a novel framework of Pneumatic Bionic Voice Prostheses as an electronic adaptation of the Pneumatic Artificial Larynx (PAL) device. The PAL is a non-invasive mechanical voice source, driven exclusively by respiration with an exceptionally high voice quality, comparable to the existing gold standard of Tracheoesophageal (TE) voice prosthesis. Following PAL design closely as the reference, Pneumatic Bionic Voice Prostheses seem to have a strong potential to substitute the existing gold standard by generating a similar voice quality while remaining non-invasive and non-surgical. This paper designs the first Pneumatic Bionic Voice prosthesis and evaluates its onset and offset control against the PAL device through pre-clinical trials on one laryngectomy patient. The evaluation on a database of more than five hours of continuous/isolated speech recordings shows a close match between the onset/offset control of the Pneumatic Bionic Voice and the PAL with an accuracy of 98.45 ±0.54%. When implemented in real-time, the Pneumatic Bionic Voice prosthesis controller has an average onset/offset delay of 10 milliseconds compared to the PAL. Hence it addresses a major disadvantage of previous electronic voice prostheses, including myoelectric Bionic Voice, in meeting the short time-frames of controlling the onset/offset of the voice in continuous speech. PMID:29466455

  18. Information transfer in verbal presentations at scientific meetings

    NASA Astrophysics Data System (ADS)

    Flinn, Edward A.

    The purpose of this note is to suggest a quantitative approach to deciding how much time to give a speaker at a scientific meeting. The elementary procedure is to use the preacher's rule of thumb that no souls are saved after the first 20 minutes. This is in qualitative agreement with the proverb that one cannot listen to a single voice for more than an hour without going to sleep. A refinement of this crude approach can be made by considering the situation from the point of view of a linear physical system with an input, a transfer function, and an output. We attempt here to derive an optimum speaking time through these considerations.

  19. Emotion Perception from Face, Voice, and Touch: Comparisons and Convergence

    PubMed Central

    Schirmer, Annett; Adolphs, Ralph

    2017-01-01

    Historically, research on emotion perception has focused on facial expressions, and findings from this modality have come to dominate our thinking about other modalities. Here, we examine emotion perception through a wider lens by comparing facial with vocal and tactile processing. We review stimulus characteristics and ensuing behavioral and brain responses, and show that audition and touch do not simply duplicate visual mechanisms. Each modality provides a distinct input channel and engages partly non-overlapping neuroanatomical systems with different processing specializations (e.g., specific emotions versus affect). Moreover, processing of signals across the different modalities converges, first into multi- and later into amodal representations that enable holistic emotion judgments. PMID:28173998

  20. Comparative speaking, shouting and singing voice range profile measurement: physiological and pathological aspects.

    PubMed

    Hacki, T

    1996-01-01

    The Voice Range Profile (VRP) measurement offers a method for the investigation of voice modalities i.e. speaking voice, shouting voice and singing voice in their mutual pitch and intensity relations. The parameters FO and SPL are evaluated by means of automatic pitch and SPL measurements from (1) sustained phonation /a:/ in the speaker's natural pitch and intensity range, (2) the continuous speaking voice beginning with Pianissimo up to Fortissimo, (3) the shouting voice. Vocal intensity is plotted vertically, vocal pitch horizontally. The displays of the vocal intensity versus fundamental frequency are defined as singing voice range profile (VRP), speaking VRP and shouting VRP. The VRPs are superimposed on the same plot. Their form, their shape and their position to each other are analysed. The physiological relationships between the VRPs of the different voice modalities to each other are defined. The pathological relationships between the VRPs (i.e. reduction, shifting) give information about etiology and pathomechanism of voice disorders.

  1. Voice activity and participation profile: assessing the impact of voice disorders on daily activities.

    PubMed

    Ma, E P; Yiu, E M

    2001-06-01

    Traditional clinical voice evaluation focuses primarily on the severity of voice impairment, with little emphasis on the impact of voice disorders on the individual's quality of life. This study reports the development of a 28-item assessment tool that evaluates the perception of voice problem, activity limitation, and participation restriction using the International Classification of Impairments, Disabilities and Handicaps-2 Beta-1 concept (World Health Organization, 1997). The questionnaire was administered to 40 subjects with dysphonia and 40 control subjects with normal voices. Results showed that the dysphonic group reported significantly more severe voice problems, limitation in daily voice activities, and restricted participation in these activities than the control group. The study also showed that the perception of a voice problem by the dysphonic subjects correlated positively with the perception of limitation in voice activities and restricted participation. However, the self-perceived voice problem had little correlation with the degree of voice-quality impairment measured acoustically and perceptually by speech pathologists. The data also showed that the aggregate scores of activity limitation and participation restriction were positively correlated, and the extent of activity limitation and participation restriction was similar in all except the job area. These findings highlight the importance of identifying and quantifying the impact of dysphonia on the individual's quality of life in the clinical management of voice disorders.

  2. Benefits for Voice Learning Caused by Concurrent Faces Develop over Time.

    PubMed

    Zäske, Romi; Mühl, Constanze; Schweinberger, Stefan R

    2015-01-01

    Recognition of personally familiar voices benefits from the concurrent presentation of the corresponding speakers' faces. This effect of audiovisual integration is most pronounced for voices combined with dynamic articulating faces. However, it is unclear if learning unfamiliar voices also benefits from audiovisual face-voice integration or, alternatively, is hampered by attentional capture of faces, i.e., "face-overshadowing". In six study-test cycles we compared the recognition of newly-learned voices following unimodal voice learning vs. bimodal face-voice learning with either static (Exp. 1) or dynamic articulating faces (Exp. 2). Voice recognition accuracies significantly increased for bimodal learning across study-test cycles while remaining stable for unimodal learning, as reflected in numerical costs of bimodal relative to unimodal voice learning in the first two study-test cycles and benefits in the last two cycles. This was independent of whether faces were static images (Exp. 1) or dynamic videos (Exp. 2). In both experiments, slower reaction times to voices previously studied with faces compared to voices only may result from visual search for faces during memory retrieval. A general decrease of reaction times across study-test cycles suggests facilitated recognition with more speaker repetitions. Overall, our data suggest two simultaneous and opposing mechanisms during bimodal face-voice learning: while attentional capture of faces may initially impede voice learning, audiovisual integration may facilitate it thereafter.

  3. Long-term average spectrum in screening of voice quality in speech: untrained male university students.

    PubMed

    Leino, Timo

    2009-11-01

    Voice quality has mainly been studied in trained speakers, singers, and dysphonic patients. Few studies have concerned ordinary untrained university students' voices. In light of earlier studies of professional voice users, it was hypothesized that good, poor, and intermediate voices would be distinguishable on the basis of long-term average spectrum characteristics. In the present study, voice quality of 50 Finnish vocally untrained male university students was studied perceptually and using long-term average spectrum analysis of text reading samples of one minute duration. Equivalent sound level (Leq) of text reading was also measured. According to the results, the good and ordinary voices differed from the poor ones in their relatively higher sound level in the frequency range of 1-3 kHz and a prominent peak at 3-4 kHz. Good voices, however, did not differ from the ordinary voices in terms of the characteristics of the long-term average spectrum (LTAS). The strength of the peak at 3-4 kHz and the voice-quality scores correlated weakly but significantly. Voice quality and alpha ratio (level difference above and below 1 kHz) correlated likewise. Leq was significantly higher in the students with good and ordinary voices than in those with poor voices. The connections between Leq, voice quality, and the formation of the peak at 3-4 kHz warrant further studies.

  4. Similar representations of emotions across faces and voices.

    PubMed

    Kuhn, Lisa Katharina; Wydell, Taeko; Lavan, Nadine; McGettigan, Carolyn; Garrido, Lúcia

    2017-09-01

    [Correction Notice: An Erratum for this article was reported in Vol 17(6) of Emotion (see record 2017-18585-001). In the article, the copyright attribution was incorrectly listed and the Creative Commons CC-BY license disclaimer was incorrectly omitted from the author note. The correct copyright is "© 2017 The Author(s)" and the omitted disclaimer is below. All versions of this article have been corrected. "This article has been published under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Copyright for this article is retained by the author(s). Author(s) grant(s) the American Psychological Association the exclusive right to publish the article and identify itself as the original publisher."] Emotions are a vital component of social communication, carried across a range of modalities and via different perceptual signals such as specific muscle contractions in the face and in the upper respiratory system. Previous studies have found that emotion recognition impairments after brain damage depend on the modality of presentation: recognition from faces may be impaired whereas recognition from voices remains preserved, and vice versa. On the other hand, there is also evidence for shared neural activation during emotion processing in both modalities. In a behavioral study, we investigated whether there are shared representations in the recognition of emotions from faces and voices. We used a within-subjects design in which participants rated the intensity of facial expressions and nonverbal vocalizations for each of the 6 basic emotion labels. For each participant and each modality, we then computed a representation matrix with the intensity ratings of each emotion. These matrices allowed us to examine the patterns of confusions between emotions and to characterize the representations of emotions within each modality. We then compared the representations across modalities by computing the correlations of the representation matrices across faces and voices. We found highly correlated matrices across modalities, which suggest similar representations of emotions across faces and voices. We also showed that these results could not be explained by commonalities between low-level visual and acoustic properties of the stimuli. We thus propose that there are similar or shared coding mechanisms for emotions which may act independently of modality, despite their distinct perceptual inputs. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  5. Effective seat-to-head transmissibility in whole-body vibration: Effects of posture and arm position

    NASA Astrophysics Data System (ADS)

    Rahmatalla, Salam; DeShaw, Jonathan

    2011-12-01

    Seat-to-head transmissibility is a biomechanical measure that has been widely used for many decades to evaluate seat dynamics and human response to vibration. Traditionally, transmissibility has been used to correlate single-input or multiple-input with single-output motion; it has not been effectively used for multiple-input and multiple-output scenarios due to the complexity of dealing with the coupled motions caused by the cross-axis effect. This work presents a novel approach to use transmissibility effectively for single- and multiple-input and multiple-output whole-body vibrations. In this regard, the full transmissibility matrix is transformed into a single graph, such as those for single-input and single-output motions. Singular value decomposition and maximum distortion energy theory were used to achieve the latter goal. Seat-to-head transmissibility matrices for single-input/multiple-output in the fore-aft direction, single-input/multiple-output in the vertical direction, and multiple-input/multiple-output directions are investigated in this work. A total of ten subjects participated in this study. Discrete frequencies of 0.5-16 Hz were used for the fore-aft direction using supported and unsupported back postures. Random ride files from a dozer machine were used for the vertical and multiple-axis scenarios considering two arm postures: using the armrests or grasping the steering wheel. For single-input/multiple-output, the results showed that the proposed method was very effective in showing the frequencies where the transmissibility is mostly sensitive for the two sitting postures and two arm positions. For multiple-input/multiple-output, the results showed that the proposed effective transmissibility indicated higher values for the armrest-supported posture than for the steering-wheel-supported posture.

  6. Integrated Nuclear and Conventional Theater Warfare Simulation (INWARS) Documentation. Part IV. User’s Manual Component. Volume III. EAD C2I Inputs.

    DTIC Science & Technology

    1980-02-08

    hours 0 Input Format: Integer b. Creatina Rescource Allocation Blocks The creation of a specific resource allocation block as a directive component is...is directed. 0 Range: N/A . Input Format: INT/NUC/CHM b. Creatina Employment Packages An employment package block has the structure portrayed in Figure

  7. The Effect of Residual Acoustic Hearing and Adaptation to Uncertainty on Speech Perception in Cochlear Implant Users: Evidence from Eye-Tracking

    PubMed Central

    McMurray, Bob; Farris-Trimble, Ashley; Seedorff, Michael; Rigler, Hannah

    2015-01-01

    Objectives While outcomes with cochlear implants (CIs) are generally good, performance can be fragile. The authors examined two factors that are crucial for good CI performance. First, while there is a clear benefit for adding residual acoustic hearing to CI stimulation (typically in low frequencies), it is unclear whether this contributes directly to phonetic categorization. Thus, the authors examined perception of voicing (which uses low-frequency acoustic cues) and fricative place of articulation (s/ʃ, which does not) in CI users with and without residual acoustic hearing. Second, in speech categorization experiments, CI users typically show shallower identification functions. These are typically interpreted as deriving from noisy encoding of the signal. However, psycholinguistic work suggests shallow slopes may also be a useful way to adapt to uncertainty. The authors thus employed an eye-tracking paradigm to examine this in CI users. Design Participants were 30 CI users (with a variety of configurations) and 22 age-matched normal hearing (NH) controls. Participants heard tokens from six b/p and six s/ʃ continua (eight steps) spanning real words (e.g., beach/peach, sip/ship). Participants selected the picture corresponding to the word they heard from a screen containing four items (a b-, p-, s- and ʃ-initial item). Eye movements to each object were monitored as a measure of how strongly they were considering each interpretation in the moments leading up to their final percept. Results Mouse-click results (analogous to phoneme identification) for voicing showed a shallower slope for CI users than NH listeners, but no differences between CI users with and without residual acoustic hearing. For fricatives, CI users also showed a shallower slope, but unexpectedly, acoustic + electric listeners showed an even shallower slope. Eye movements showed a gradient response to fine-grained acoustic differences for all listeners. Even considering only trials in which a participant clicked “b” (for example), and accounting for variation in the category boundary, participants made more looks to the competitor (“p”) as the voice onset time neared the boundary. CI users showed a similar pattern, but looked to the competitor more than NH listeners, and this was not different at different continuum steps. Conclusion Residual acoustic hearing did not improve voicing categorization suggesting it may not help identify these phonetic cues. The fact that acoustic + electric users showed poorer performance on fricatives was unexpected as they usually show a benefit in standardized perception measures, and as sibilants contain little energy in the low-frequency (acoustic) range. The authors hypothesize that these listeners may over-weight acoustic input, and have problems when this is not available (in fricatives). Thus, the benefit (or cost) of acoustic hearing for phonetic categorization may be complex. Eye movements suggest that in both CI and NH listeners, phoneme categorization is not a process of mapping continuous cues to discrete categories. Rather listeners preserve gradiency as a way to deal with uncertainty. CI listeners appear to adapt to their implant (in part) by amplifying competitor activation to preserve their flexibility in the face of potential misperceptions. PMID:26317298

  8. Randomized Controlled Trial of Supplemental Augmentative and Alternative Communication versus Voice Rest Alone after Phonomicrosurgery

    PubMed Central

    Rousseau, Bernard; Gutmann, Michelle L.; Mau, I-fan Theodore; Francis, David O.; Johnson, Jeffrey P.; Novaleski, Carolyn K.; Vinson, Kimberly N.; Garrett, C. Gaelyn

    2015-01-01

    Objective This randomized trial investigated voice rest and supplemental text-to-speech communication versus voice rest alone on visual analog scale measures of communication effectiveness and magnitude of voice use. Study Design Randomized clinical trial. Setting Multicenter outpatient voice clinics. Subjects Thirty-seven patients undergoing phonomicrosurgery. Methods Patients undergoing phonomicrosurgery were randomized to voice rest and supplemental text-to-speech communication or voice rest alone. The primary outcome measure was the impact of voice rest on ability to communicate effectively over a seven-day period. Pre- and post-operative magnitude of voice use was also measured as an observational outcome. Results Patients randomized to voice rest and supplemental text-to-speech communication reported higher median communication effectiveness on each post-operative day compared to those randomized to voice rest alone, with significantly higher median communication effectiveness on post-operative day 3 (p = 0.03) and 5 (p = 0.01). Magnitude of voice use did not differ on any pre-operative (p > 0.05) or post-operative day (p > 0.05), nor did patients significantly decrease voice use as the surgery date approached (p > 0.05). However, there was a significant reduction in median voice use pre- to post-operatively across patients (p < 0.001) with median voice use ranging from 0–3 throughout the post-operative week. Conclusion Supplemental text-to-speech communication increased patient perceived communication effectiveness on post-operative days 3 and 5 over voice rest alone. With the prevalence of smartphones and the widespread use of text messaging, supplemental text-to-speech communication may provide an accessible and cost-effective communication option for patients on vocal restrictions. PMID:25605690

  9. Influence of consonant voicing characteristics on sentence production in abductor versus adductor spasmodic dysphonia.

    PubMed

    Cannito, Michael P; Chorna, Lesya B; Kahane, Joel C; Dworkin, James P

    2014-05-01

    This study evaluated the hypotheses that sentence production by speakers with adductor (AD) and abductor (AB) spasmodic dysphonia (SD) may be differentially influenced by consonant voicing and manner features, in comparison with healthy, matched, nondysphonic controls. This was a prospective, single blind study, using a between-groups, repeated measures design for the independent variables of perceived voice quality and sentence duration. Sixteen subjects with ADSD and 10 subjects with ABSD, as well as 26 matched healthy controls produced four short, simple sentences that were systematically loaded with voiced or voiceless consonants of either obstruant or continuant manner categories. Experienced voice clinicians, who were "blind" as to speakers' group affixations, used visual analog scaling to judge the overall voice quality of each sentence. Acoustic sentence durations were also measured. Speakers with ABSD or ADSD demonstrated significantly poorer than normal voice quality on all sentences. Speakers with ABSD exhibited longer than normal duration for voiceless consonant sentences. Speakers with ADSD had poorer voice quality for voiced than for voiceless consonant sentences. Speakers with ABSD had longer durations for voiceless than for voiced consonant sentences. The two subtypes of SD exhibit differential performance on the basis of consonant voicing in short, simple sentences; however, each subgroup manifested voicing-related differences on a different variable (voice quality vs sentence duration). Findings suggest different underlying pathophysiological mechanisms for ABSD and ADSD. Findings also support inclusion of short, simple sentences containing voiced or voiceless consonants as part of the diagnostic protocol for SD, with measurement of sentence duration in addition to judments of voice quality severity. Copyright © 2014 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  10. Hearing the Unheard: An Interdisciplinary, Mixed Methodology Study of Women's Experiences of Hearing Voices (Auditory Verbal Hallucinations).

    PubMed

    McCarthy-Jones, Simon; Castro Romero, Maria; McCarthy-Jones, Roseline; Dillon, Jacqui; Cooper-Rompato, Christine; Kieran, Kathryn; Kaufman, Milissa; Blackman, Lisa

    2015-01-01

    This paper explores the experiences of women who "hear voices" (auditory verbal hallucinations). We begin by examining historical understandings of women hearing voices, showing these have been driven by androcentric theories of how women's bodies functioned leading to women being viewed as requiring their voices be interpreted by men. We show the twentieth century was associated with recognition that the mental violation of women's minds (represented by some voice-hearing) was often a consequence of the physical violation of women's bodies. We next report the results of a qualitative study into voice-hearing women's experiences (n = 8). This found similarities between women's relationships with their voices and their relationships with others and the wider social context. Finally, we present results from a quantitative study comparing voice-hearing in women (n = 65) and men (n = 132) in a psychiatric setting. Women were more likely than men to have certain forms of voice-hearing (voices conversing) and to have antecedent events of trauma, physical illness, and relationship problems. Voices identified as female may have more positive affect than male voices. We conclude that women voice-hearers have and continue to face specific challenges necessitating research and activism, and hope this paper will act as a stimulus to such work.

  11. The effectiveness of a voice treatment approach for teachers with self-reported voice problems.

    PubMed

    Gillivan-Murphy, Patricia; Drinnan, Michael J; O'Dwyer, Tadhg P; Ridha, Hayder; Carding, Paul

    2006-09-01

    Teachers are considered the professional group most at risk of developing voice-problems, but limited treatment effectiveness evidence exists. We studied prospectively the effectiveness of a 6-week combined treatment approach using vocal function exercises (VFEs) and vocal hygiene (VH) education with 20 teachers with self-reported voice problems. Twenty subjects were randomly assigned to a no-treatment control (n = 11) and a treatment group (n = 9). Fibreoptic endoscopic evaluation was carried out on all subjects before randomization. Two self-report voice outcome measures were used: the Voice-Related Quality of Life (VRQOL) and the Voice Symptom Severity Scale (VoiSS). A Voice Care Knowledge Visual Analogue Scale (VAS), developed specifically for the study, was also used to evaluate change in selected voice knowledge areas. A Student unpaired t test revealed a statistically significant (P < 0.05) improvement in the treatment group as measured by the VoiSS. There was not a significant improvement in the treatment group as measured by the V-RQOL. The difference in voice care knowledge areas was also significant for the treatment group (P < 0.05). This study suggests that a voice treatment approach of VFEs and VH education improved self-reported voice symptoms and voice care knowledge in a group of teachers.

  12. Relationship between Activity Noise, Voice Parameters, and Voice Symptoms among Female Teachers.

    PubMed

    Pirilä, Sirpa; Pirilä, Paula; Ansamaa, Terhi; Yliherva, Anneli; Sonning, Samuel; Rantala, Leena

    2017-01-01

    Our interest was in how teachers' voices behave during the delivery of lessons in core subjects (e.g., mathematics, science, etc.). We sought to evaluate the relationship between voice sound pressure level (SPL), vocal fundamental frequency (F0), voice symptoms, activity noise, and differences therein during the first and the last lessons in core subjects of the day. The participants were 24 female elementary school teachers. Voice symptoms were evaluated by questionnaire. The data were recorded on 2 portable voice accumulators (VoxLog) from the first and last lessons of the day. The versions of accumulators differed by frequency weighting; therefore, the analysis and the results of noise and voice SPL were treated separately: unweighted (group 1) and A-weighted (group 2). Difference in voice SPL followed difference in activity noise. F0 increased between the first and last lessons. Correlations were found between differences in the noise and the voice symptoms of tiredness and dryness. Irritating mucus was associated with high F0 during the first lesson. An apparent increase in voice loading due to the activity noise was observed during lessons in core subjects. Collaboration between specialists in voice and acoustics and teachers and pupils is needed to reduce this voice loading. © 2017 S. Karger AG, Basel.

  13. The effect of singing training on voice quality for people with quadriplegia.

    PubMed

    Tamplin, Jeanette; Baker, Felicity A; Buttifant, Mary; Berlowitz, David J

    2014-01-01

    Despite anecdotal reports of voice impairment in quadriplegia, the exact nature of these impairments is not well described in the literature. This article details objective and subjective voice assessments for people with quadriplegia at baseline and after a respiratory-targeted singing intervention. Randomized controlled trial. Twenty-four participants with quadriplegia were randomly assigned to a 12-week program of either a singing intervention or active music therapy control. Recordings of singing and speech were made at baseline, 6 weeks, 12 weeks, and 6 months postintervention. These deidentified recordings were used to measure sound pressure levels and assess voice quality using the Multidimensional Voice Profile and the Perceptual Voice Profile. Baseline voice quality data indicated deviation from normality in the areas of breathiness, strain, and roughness. A greater percentage of intervention participants moved toward more normal voice quality in terms of jitter, shimmer, and noise-to-harmonic ratio; however, the improvements failed to achieve statistical significance. Subjective and objective assessments of voice quality indicate that quadriplegia may have a detrimental effect on voice quality; in particular, causing a perception of roughness and breathiness in the voice. The results of this study suggest that singing training may have a role in ameliorating these voice impairments. Copyright © 2014 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  14. The phonatory deviation diagram: a novel objective measurement of vocal function.

    PubMed

    Madazio, Glaucya; Leão, Sylvia; Behlau, Mara

    2011-01-01

    To identify the discriminative characteristics of the phonatory deviation diagram (PDD) in rough, breathy and tense voices. One hundred and ninety-six samples of normal and dysphonic voices from adults were submitted to perceptual auditory evaluation, focusing on the predominant vocal quality and the degree of deviation. Acoustic analysis was performed with the VoxMetria (CTS Informatica). Significant differences were observed between the dysphonic and normal groups (p < 0.001), and also between the breathy and rough samples (p = 0.044) and the breathy and tense samples (p < 0.001). All normal voices were positioned in the inferior left quadrant, 45% of the rough voices in the inferior right quadrant, 52.6% of the breathy voices in the superior right quadrant and 54.3% of the tense voices in the inferior left quadrant of the PDD. In the inferior left quadrant, 93.8% of voices with no deviation were located and 72.7% of voices with mild deviation; voices with moderate deviation were distributed in the inferior and superior right quadrants, the latter ones containing the most deviant voices and 80% of voices with severe deviation. The PDD was able to discriminate normal from dysphonic voices, and the distribution was related to the type and degree of voice alteration. Copyright © 2011 S. Karger AG, Basel.

  15. Voice and choice in health care in England: understanding citizen responses to dissatisfaction.

    PubMed

    Dowding, Keith; John, Peter

    2011-01-01

    Using data from a five-year online survey the paper examines the effects of relative satisfaction with health services on individuals' voice-and-choice activity in the English public health care system. Voice is considered in three parts – individual voice (complaints), collective voice voting and participation (collective action). Exercising choice is seen in terms of complete exit (not using health care), internal exit (choosing another public service provider) and private exit (using private health care). The interaction of satisfaction and forms of voice and choice are analysed over time. Both voice and choice are correlated with dissatisfaction with those who are unhappy with the NHS more likely to privately voice and to plan to take up private health care. Those unable to choose private provision are likely to use private voice. These factors are not affected by items associated with social capital – indeed, being more trusting leads to lower voice activity.

  16. Can health insurance improve access to quality care for the Indian poor?

    PubMed

    Michielsen, Joris; Criel, Bart; Devadasan, Narayanan; Soors, Werner; Wouters, Edwin; Meulemans, Herman

    2011-08-01

    Recently, the Indian government launched health insurance schemes for the poor both to protect them from high health spending and to improve access to high-quality health services. This article aims to review the potentials of health insurance interventions in order to improve access to quality care in India based on experiences of community health insurance schemes. PubMed, Ovid MEDLINE (R), All EBM Reviews, CSA Sociological Abstracts, CSA Social Service Abstracts, EconLit, Science Direct, the ISI Web of Knowledge, Social Science Research Network and databases of research centers were searched up to September 2010. An Internet search was executed. One thousand hundred and thirty-three papers were assessed for inclusion and exclusion criteria. Twenty-five papers were selected providing information on eight schemes. A realist review was performed using Hirschman's exit-voice theory: mechanisms to improve exit strategies (financial assets and infrastructure) and strengthen patient's long voice route (quality management) and short voice route (patient pressure). All schemes use a mix of measures to improve exit strategies and the long voice route. Most mechanisms are not effective in reality. Schemes that focus on the patients' bargaining position at the patient-provider interface seem to improve access to quality care. Top-down health insurance interventions with focus on exit strategies will not work out fully in the Indian context. Government must actively facilitate the potential of CHI schemes to emancipate the target group so that they may transform from mere passive beneficiaries into active participants in their health.

  17. Voice- and swallow-related quality of life in idiopathic Parkinson's disease.

    PubMed

    van Hooren, Michel R A; Baijens, Laura W J; Vos, Rein; Pilz, Walmari; Kuijpers, Laura M F; Kremer, Bernd; Michou, Emilia

    2016-02-01

    This study explores whether changes in voice- and swallow-related QoL are associated with progression of idiopathic Parkinson's disease (IPD). Furthermore, it examines the relationship between patients' perception of both voice and swallowing disorders in IPD. Prospective clinical study, quality of life (QoL). One-hundred mentally competent IPD patients with voice and swallowing complaints were asked to answer four QoL questionnaires (Voice Handicap Index, MD Anderson Dysphagia Inventory, Visual Analog Scale [VAS] voice, and Dysphagia Severity Scale [DSS]). Differences in means for the QoL questionnaires and their subscales within Hoehn and Yahr stage groups were calculated using one-way analysis of variance. The relationship between voice- and swallow-related QoL questionnaires was determined with the Spearman correlation coefficient. Scores on both voice and swallow questionnaires suggest an overall decrease in QoL with progression of IPD. A plateau in QoL for VAS voice and the DSS was seen in the early Hoehn and Yahr stages. Finally, scores on voice-related QoL questionnaires were significantly correlated with swallow-related QoL outcomes. Voice- and swallow-related QoL decreases with progression of IPD. A significant association was found between voice- and swallow-related QoL questionnaires. Healthcare professionals can benefit from voice- and swallow-related QoL questionnaires in a multidimensional voice- or swallow-assessment protocol. The patient's perception of his/her voice and swallowing disorders and its impact on QoL in IPD should not be disregarded. 2b. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.

  18. Neural circuits underlying mother’s voice perception predict social communication abilities in children

    PubMed Central

    Abrams, Daniel A.; Chen, Tianwen; Odriozola, Paola; Cheng, Katherine M.; Baker, Amanda E.; Padmanabhan, Aarthi; Ryali, Srikanth; Kochalka, John; Feinstein, Carl; Menon, Vinod

    2016-01-01

    The human voice is a critical social cue, and listeners are extremely sensitive to the voices in their environment. One of the most salient voices in a child’s life is mother's voice: Infants discriminate their mother’s voice from the first days of life, and this stimulus is associated with guiding emotional and social function during development. Little is known regarding the functional circuits that are selectively engaged in children by biologically salient voices such as mother’s voice or whether this brain activity is related to children’s social communication abilities. We used functional MRI to measure brain activity in 24 healthy children (mean age, 10.2 y) while they attended to brief (<1 s) nonsense words produced by their biological mother and two female control voices and explored relationships between speech-evoked neural activity and social function. Compared to female control voices, mother’s voice elicited greater activity in primary auditory regions in the midbrain and cortex; voice-selective superior temporal sulcus (STS); the amygdala, which is crucial for processing of affect; nucleus accumbens and orbitofrontal cortex of the reward circuit; anterior insula and cingulate of the salience network; and a subregion of fusiform gyrus associated with face perception. The strength of brain connectivity between voice-selective STS and reward, affective, salience, memory, and face-processing regions during mother’s voice perception predicted social communication skills. Our findings provide a novel neurobiological template for investigation of typical social development as well as clinical disorders, such as autism, in which perception of biologically and socially salient voices may be impaired. PMID:27185915

  19. Neural circuits underlying mother's voice perception predict social communication abilities in children.

    PubMed

    Abrams, Daniel A; Chen, Tianwen; Odriozola, Paola; Cheng, Katherine M; Baker, Amanda E; Padmanabhan, Aarthi; Ryali, Srikanth; Kochalka, John; Feinstein, Carl; Menon, Vinod

    2016-05-31

    The human voice is a critical social cue, and listeners are extremely sensitive to the voices in their environment. One of the most salient voices in a child's life is mother's voice: Infants discriminate their mother's voice from the first days of life, and this stimulus is associated with guiding emotional and social function during development. Little is known regarding the functional circuits that are selectively engaged in children by biologically salient voices such as mother's voice or whether this brain activity is related to children's social communication abilities. We used functional MRI to measure brain activity in 24 healthy children (mean age, 10.2 y) while they attended to brief (<1 s) nonsense words produced by their biological mother and two female control voices and explored relationships between speech-evoked neural activity and social function. Compared to female control voices, mother's voice elicited greater activity in primary auditory regions in the midbrain and cortex; voice-selective superior temporal sulcus (STS); the amygdala, which is crucial for processing of affect; nucleus accumbens and orbitofrontal cortex of the reward circuit; anterior insula and cingulate of the salience network; and a subregion of fusiform gyrus associated with face perception. The strength of brain connectivity between voice-selective STS and reward, affective, salience, memory, and face-processing regions during mother's voice perception predicted social communication skills. Our findings provide a novel neurobiological template for investigation of typical social development as well as clinical disorders, such as autism, in which perception of biologically and socially salient voices may be impaired.

  20. Perceptions of Voice Teachers Regarding Students' Vocal Behaviors During Singing and Speaking.

    PubMed

    Beeman, Shellie A

    2017-01-01

    This study examined voice teachers' perceptions of their instruction of healthy singing and speaking voice techniques. An online, researcher-generated questionnaire based on the McClosky technique was administered to college/university voice teachers listed as members in the 2012-2013 College Music Society directory. A majority of participants believed there to be a relationship between the health of the singing voice and the health of the speaking voice. Participants' perception scores were the most positive for variable MBSi, the monitoring of students' vocal behaviors during singing. Perception scores for variable TVB, the teaching of healthy vocal behaviors, and variable MBSp, the monitoring of students' vocal behaviors while speaking, ranked second and third, respectively. Perception scores for variable TVB were primarily associated with participants' familiarity with voice rehabilitation techniques, gender, and familiarity with the McClosky technique. Perception scores for variable MBSi were primarily associated with participants' familiarity with voice rehabilitation techniques, gender, type of student taught, and instruction of a student with a voice disorder. Perception scores for variable MBSp were correlated with the greatest number of characteristics, including participants' familiarity with voice rehabilitation techniques, familiarity with the McClosky technique, type of student taught, years of teaching experience, and instruction of a student with a voice disorder. Voice teachers are purportedly working with injured voices and attempting to include vocal health in their instruction. Although a voice teacher is not obligated to pursue further rehabilitative training, the current study revealed a positive relationship between familiarity with specific rehabilitation techniques and vocal health. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  1. Performer's attitudes toward seeking health care for voice issues: understanding the barriers.

    PubMed

    Gilman, Marina; Merati, Albert L; Klein, Adam M; Hapner, Edie R; Johns, Michael M

    2009-03-01

    Contemporary commercial music (CCM) performers rely heavily on their voice, yet may not be aware of the importance of proactive voice care. This investigation intends to identify perceptions and barriers to seeking voice care among CCM artists. This cross-sectional observational study used a 10-item Likert-based response questionnaire to assess current perceptions regarding voice care in a population of randomly selected participants of professional CCM conference. Subjects (n=78) were queried regarding their likelihood to seek medical care for minor medical problems and specifically problems with their voice. Additional questions investigated anxiety about seeking voice care from a physician specialist, speech language pathologist, or voice coach; apprehension regarding findings of laryngeal examination, laryngeal imaging procedures; and the effect of medical insurance on the likelihood of seeking medical care. Eighty-two percent of subjects reported that their voice was a critical part of their profession; 41% stated that they were not likely to seek medical care for problems with their voice; and only 19% were reluctant to seek care for general medical problems (P<0.001). Anxiety about seeking a clinician regarding their voice was not a deterrent. Most importantly, 39% of subjects do not seek medical attention for their voice problems due to medical insurance coverage. The CCM artists are less likely to seek medical care for voice problems compared with general medical problems. Availability of medical insurance may be a factor. Availability of affordable voice care and education about the importance of voice care is needed in this population of vocal performers.

  2. Hidden student voice: A curriculum of a middle school science class heard through currere

    NASA Astrophysics Data System (ADS)

    Crooks, Kathleen Schwartz

    Students have their own lenses through which they view school science and the students' views are often left out of educational conversations which directly affect the students themselves. Pinar's (2004) definition of curriculum as a 'complicated conversation' implies that the class' voice is important, as important as the teacher's voice, to the classroom conversation. If the class' voice is vital to classroom conversations, then the class, consisting of all its students, must be allowed to both speak and be heard. Through a qualitative case study, whereby the case is defined as a particular middle school science class, this research attempts to hear the 'complicated conversation' of this middle school science class, using currere as a framework. Currere suggests that one's personal relationship to the world, including one's memories, hopes, and dreams, should be the crux of education, rather than education being primarily the study of facts, concepts, and needs determined by an 'other'. Focus group interviews were used to access the class' currere: the class' lived experiences of science, future dreams of science, and present experiences of science, which was synthesized into a new understanding of the present which offered the class the opportunity to be fully educated. The interview data was enriched through long-term observation in this middle school science classroom. Analysis of the data collected suggests that a middle school science class has rich science stories which may provide insights into ways to engage more students in science. Also, listening to the voice of a science class may provide insight into discussions about science education and understandings into the decline in student interest in science during secondary school. Implications from this research suggest that school science may be more engaging for this middle school class if it offers inquiry-based activities and allows opportunities for student-led research. In addition, specialized academic and career advice in early middle school may be able to capitalize on this class' positive perspective toward science. Further research may include using currere to hear the voices of middle school science classes with more diverse demographic qualities.

  3. Voice similarity in identical twins.

    PubMed

    Van Gysel, W D; Vercammen, J; Debruyne, F

    2001-01-01

    If people are asked to discriminate visually the two individuals of a monozygotic twin (MT), they mostly get into trouble. Does this problem also exist when listening to twin voices? Twenty female and 10 male MT voices were randomly assembled with one "strange" voice to get voice trios. The listeners (10 female students in Speech and Language Pathology) were asked to label the twins (voices 1-2, 1-3 or 2-3) in two conditions: two standard sentences read aloud and a 2.5-second midsection of a sustained /a/. The proportion correctly labelled twins was for female voices 82% and 63% and for male voices 74% and 52% for the sentences and the sustained /a/ respectively, both being significantly greater than chance (33%). The acoustic analysis revealed a high intra-twin correlation for the speaking fundamental frequency (SFF) of the sentences and the fundamental frequency (F0) of the sustained /a/. So the voice pitch could have been a useful characteristic in the perceptual identification of the twins. We conclude that there is a greater perceptual resemblance between the voices of identical twins than between voices without genetic relationship. The identification however is not perfect. The voice pitch possibly contributes to the correct twin identifications.

  4. [Applicability of voice acoustic analysis with vocal loading testto diagnostics of occupational voice diseases].

    PubMed

    Niebudek-Bogusz, Ewa; Sliwińska-Kowalska, Mariola

    2006-01-01

    An assessment of the vocal system, as a part of the medical certification of occupational diseases, should be objective and reliable. Therefore, interest in the method of acoustic voice analysis enabling objective assessment of voice parameters is still growing. The aim of the present study was to evaluate the applicability of acoustic analysis with vocal loading test to the diagnostics of occupational voice disorders. The results of acoustic voice analysis were compared using IRIS software for phoniatrics, before and after a 30-min vocal loading test in 35 female teachers with diagnosed occupational voice disorders (group I) and in 31 female teachers with functional dysphonia (group II). In group I, vocal effort produced significant abnormalities in voice acoustic parameters, compared to group II. These included significantly increased mean fundamental frequency (Fo) value (by 11 Hz) and worsened jitter, shimmer and NHR parameters. Also, the percentage of subjects showing abnormalities in voice acoustic analysis was higher in this group. Conducting voice acoustic analysis before and after the vocal loading test makes it possible to objectively confirm irreversible voice impairments in persons with work-related pathologies of the larynx, which is essential for medical certification of occupational voice diseases.

  5. Skilled Voices?: Reflections on Political Participation and Education in Austria. OECD Education Working Papers, No. 11

    ERIC Educational Resources Information Center

    Walter, Florian; Rosenberger, Sieglinde

    2007-01-01

    This study, part of OECD/CERI's project on Measuring the Social Outcomes of Learning, investigates the relationship between educational attainment and political participation in Austria. First, a model based on various theoretical considerations is introduced. This incorporates direct educational effects as well as indirect effects that occur…

  6. 78 FR 67025 - Domestic Requests for Broadcasting Board of Governors Program Materials

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-11-08

    ... copyrighted materials. List of Subjects in 22 CFR Part 502 Broadcasting, Foreign relations, News media, Public... Agency program materials should be directed to: (a) The Voice of America Office of Public Relations for... from members of the public, organizations, and media, for program materials disseminated by BBG abroad...

  7. 75 FR 44855 - Migratory Bird Hunting; Proposed Frameworks for Early-Season Migratory Bird Hunting Regulations...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-07-29

    ... Duck Seasons iii. Black Ducks iv. Canvasbacks v. Pintails vi. Scaup vii. Mottled Ducks viii. Wood Ducks... with specific harvest strategies (canvasbacks, pintails, black ducks, and scaup), those strategies will... active voice to address readers directly; (c) Use clear language rather than jargon; (d) Be divided into...

  8. Voices of Black Women Superintendents' Experiences

    ERIC Educational Resources Information Center

    Cox, Gloria

    2017-01-01

    The shortage of African American leadership in business, science, law, medicine or just in public school districts can be directly linked to several factors, including a shortage of African American teachers who will enter the leadership pipeline in both urban and suburban settings. This study uses a narrative approach to explore the experiences…

  9. Comparison of CDMA and FDMA for the MobileStar(sm) system

    NASA Technical Reports Server (NTRS)

    Jacobs, I. M.; Gilhousen, K. S.; Weaver, L. A.; Renshaw, K.; Murphy, T.

    1988-01-01

    Spread-spectrum code division multiple access (CDMA) and single channel per carrier frequency division multiple access (FDMA) systems are compared for spectrum efficiency. CDMA is shown to have greater maximum throughput than FDMA for the MobileStar(sm) system which uses digital voice activated carriers and directive circularly polarized satellite antennas.

  10. Democracy--Unleashing the Power of "We"

    ERIC Educational Resources Information Center

    Heaney, Tom

    2015-01-01

    An obvious goal of adult learners is to find their own voice, to be heard in rational discourse with their peers, and to gain control over the day-to-day decisions that affect their lives. This chapter asks how doctoral students can be partners with faculty in charting the direction of their academic pursuits.

  11. A Quiet Voice

    ERIC Educational Resources Information Center

    Schrader, Teri

    2009-01-01

    The Common Principles have been at the very center of the author's professional practice. When she first read Ted Sizer's writing and learned about the Coalition of Essential Schools, she felt as though he was talking directly to her. Not only did every word of the then nine Common Principles make sense, but after reading Sizer's work, her own…

  12. PERCEPTUAL SYSTEMS IN READING--THE PREDICTION OF A TEMPORAL EYE-VOICE SPAN CONSTANT. PAPER.

    ERIC Educational Resources Information Center

    GEYER, JOHN JACOB

    A STUDY WAS CONDUCTED TO DELINEATE HOW PERCEPTION OCCURS DURING ORAL READING. FROM AN ANALYSIS OF CLASSICAL AND MODERN RESEARCH, A HEURISTIC MODEL WAS CONSTRUCTED WHICH DELINEATED THE DIRECTLY INTERACTING SYSTEMS POSTULATED AS FUNCTIONING DURING ORAL READING. THE MODEL AS OUTLINED WAS DIFFERENTIATED LOGICALLY INTO THREE MAJOR PROCESSING…

  13. He Who Hesitates Fears Cost

    ERIC Educational Resources Information Center

    Korzeniowski, Paul

    2009-01-01

    Voice over IP (VoIP) has been infiltrating campus networks, but more like stray weeds in an unattended garden than like a well-planned crop. Trouble is, in most instances, moving directly from a PBX or Centrex service to VoIP represents a shift too costly and dramatic for many academic institutions to undertake. Instead, schools have been…

  14. Coconut Allergy Revisited

    PubMed Central

    Anagnostou, Katherine

    2017-01-01

    Despite concerns voiced often by food-allergic patients, allergy to coconut is rare, not directly associated with nut allergy and few cases are reported so far in the literature. We present an interesting case of coconut allergy in a child that was previously tolerant to coconut and regularly exposed via both the skin and gastrointestinal route. PMID:28961189

  15. Using Student Voice to Respond to Middle School Bullying: A Student Leadership Approach

    ERIC Educational Resources Information Center

    Shriberg, David; Brooks, Keeshawna; Jenkins, Kisha; Immen, Jennifer; Sutter, Caroline; Cronin, Karen

    2017-01-01

    Bullying prevention and intervention are ongoing challenges for all educators, school psychologists included. A lack of research exists regarding the potential role of middle school students as direct actors in bullying prevention and intervention. This article describes a novel student leadership group for seventh graders in which the primary…

  16. English and Socio-Economic Disadvantage: Learner Voices from Rural Bangladesh

    ERIC Educational Resources Information Center

    Hamid, M. Obaidul; Baldauf, Richard B., Jr.

    2011-01-01

    L2 education research has shown immense interest in learners and their views of L2 learning. Nevertheless, the different directions of learner-focused research have been inadequate in highlighting learners' learning experiences in relation to their social backgrounds, particularly in the developing world. Drawing on the first author's PhD…

  17. 47 CFR 80.1077 - Frequencies.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... System: Alerting: 406.0-406.1 EPIRBs 406.0-406.1 MHz (Earth-to-space).1544-1545 MHz (space-to-Earth). INMARSAT Ship Earth Stations capable of voice and/or direct printing 1626.5-1645.5 MHz (Earth-to-space... safety communications and calling: Satellite 1530-1544 MHz (space-to-Earth) and 1626.5-1645.5 MHz (Earth...

  18. 47 CFR 80.1077 - Frequencies.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... System: Alerting: 406.0-406.1 EPIRBs 406.0-406.1 MHz (Earth-to-space).1544-1545 MHz (space-to-Earth). INMARSAT Ship Earth Stations capable of voice and/or direct printing 1626.5-1645.5 MHz (Earth-to-space... safety communications and calling: Satellite 1530-1544 MHz (space-to-Earth) and 1626.5-1645.5 MHz (Earth...

  19. 47 CFR 80.1077 - Frequencies.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... System: Alerting: 406.0-406.1 EPIRBs 406.0-406.1 MHz (Earth-to-space).1544-1545 MHz (space-to-Earth). INMARSAT Ship Earth Stations capable of voice and/or direct printing 1626.5-1645.5 MHz (Earth-to-space... safety communications and calling: Satellite 1530-1544 MHz (space-to-Earth) and 1626.5-1645.5 MHz (Earth...

  20. Photovoice for Healthy Relationships: Community-Based Participatory HIV Prevention in a Rural American Indian Community

    ERIC Educational Resources Information Center

    Markus, Susan F.

    2012-01-01

    This article provides an example of a culturally responsive, community-based project for addressing social determinants of health in rural American Indian (AI) communities through: 1) empowering youth and community voices to set directions for HIV, sexually transmitted infections, and unintended pregnancy prevention and education efforts; 2) using…

  1. Research and Clinical Center for Child Development Annual Report, 1999-2000, No. 23.

    ERIC Educational Resources Information Center

    Chen, Shing-Jen, Ed.; Fujino, Yuki, Ed.

    This annual report presents several articles related to the work of the Clinical Center for Child Development at Hokkaido University in Sapporo, Japan. The articles are: (1) "Intrinsic Musicality: Rhythm and Prosody in Infant-Directed Voices" (Niki Powers); (2) "Movable Cognitive Studies with a Portable, Telemetric Near-Infrared…

  2. Listening to the Voice of the Customer.

    ERIC Educational Resources Information Center

    Schauerman, Sam; And Others

    One of the major tenets of Total Quality Management (TQM) is that organizations need to adopt a strong customer focus. At El Camino College (ECC) in Torrance, California, a matrix was developed to identify and describe ECC's direct and indirect internal and external customers. ECC then applied Quality Function Deployment (QFD), a strategic tool…

  3. Civic Journalism and Nonelite Sourcing: Making Routine Newswork of Community Connectedness.

    ERIC Educational Resources Information Center

    Massey, Brian L.

    1998-01-01

    Compares the number of "average" citizens brought into the news in three newspapers. Finds nonelite information sources in numerical parity with elite sources in a civic-journalism newspaper, but finds the frequency and directness of their news voices largely unchanged. Finds that routine civic journalism did more to tone down elites'…

  4. Listening to Children with Communication Impairment Talking through Their Drawings

    ERIC Educational Resources Information Center

    Holliday, Erin L.; Harrison, Linda J.; McLeod, Sharynne

    2009-01-01

    Including children as research participants is an important new direction in early childhood research. However, it is rare for such studies to include the voices of children with significant communication impairment. This article suggests that drawing may be an appropriate non-verbal method for "listening" to these children's ideas and recording…

  5. 40 CFR 267.34 - When must personnel have access to communication equipment or an alarm system?

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... to an internal alarm or emergency communication device, either directly or through visual or voice... communication equipment or an alarm system? 267.34 Section 267.34 Protection of Environment ENVIRONMENTAL... have access to communication equipment or an alarm system? (a) Whenever hazardous waste is being poured...

  6. Adult Latino College Students: Experiencias y la Educacion

    ERIC Educational Resources Information Center

    Garza, Ana Lisa

    2011-01-01

    The study aimed to gain a better understanding of the learning experiences of adult Latino college students, as described directly in their own voices. The study was guided by two research questions: RQ1: "How do adult Latinos describe their undergraduate college learning experiences?" and RQ2: "How do culture, gender, and ethnic…

  7. Experimental and theoretical identification of a four- acoustic-inputs/two-vibration-outputs hearing system

    NASA Astrophysics Data System (ADS)

    Balaji, P. A.

    1999-07-01

    A cricket's ear is a directional acoustic sensor. It has a remarkable level of sensitivity to the direction of sound propagation in a narrow frequency bandwidth of 4-5 KHz. Because of its complexity, the directional sensitivity has long intrigued researchers. The cricket's ear is a four-acoustic-inputs/two-vibration-outputs system. In this dissertation, this system is examined in depth, both experimentally and theoretically, with a primary goal to understand the mechanics involved in directional hearing. Experimental identification of the system is done by using random signal processing techniques. Theoretical identification of the system is accomplished by analyzing sound transmission through complex trachea of the ear. Finally, a description of how the cricket achieves directional hearing sensitivity is proposed. The fundamental principle involved in directional heating of the cricket has been utilized to design a device to obtain a directional signal from non- directional inputs.

  8. Retinal Origin of Direction Selectivity in the Superior Colliculus

    PubMed Central

    Shi, Xuefeng; Barchini, Jad; Ledesma, Hector Acaron; Koren, David; Jin, Yanjiao; Liu, Xiaorong; Wei, Wei; Cang, Jianhua

    2017-01-01

    Detecting visual features in the environment such as motion direction is crucial for survival. The circuit mechanisms that give rise to direction selectivity in a major visual center, the superior colliculus (SC), are entirely unknown. Here, we optogenetically isolate the retinal inputs that individual direction-selective SC neurons receive and find that they are already selective as a result of precisely converging inputs from similarly-tuned retinal ganglion cells. The direction selective retinal input is linearly amplified by the intracollicular circuits without changing its preferred direction or level of selectivity. Finally, using 2-photon calcium imaging, we show that SC direction selectivity is dramatically reduced in transgenic mice that have decreased retinal selectivity. Together, our studies demonstrate a retinal origin of direction selectivity in the SC, and reveal a central visual deficit as a consequence of altered feature selectivity in the retina. PMID:28192394

  9. Speaker's voice as a memory cue.

    PubMed

    Campeanu, Sandra; Craik, Fergus I M; Alain, Claude

    2015-02-01

    Speaker's voice occupies a central role as the cornerstone of auditory social interaction. Here, we review the evidence suggesting that speaker's voice constitutes an integral context cue in auditory memory. Investigation into the nature of voice representation as a memory cue is essential to understanding auditory memory and the neural correlates which underlie it. Evidence from behavioral and electrophysiological studies suggest that while specific voice reinstatement (i.e., same speaker) often appears to facilitate word memory even without attention to voice at study, the presence of a partial benefit of similar voices between study and test is less clear. In terms of explicit memory experiments utilizing unfamiliar voices, encoding methods appear to play a pivotal role. Voice congruency effects have been found when voice is specifically attended at study (i.e., when relatively shallow, perceptual encoding takes place). These behavioral findings coincide with neural indices of memory performance such as the parietal old/new recollection effect and the late right frontal effect. The former distinguishes between correctly identified old words and correctly identified new words, and reflects voice congruency only when voice is attended at study. Characterization of the latter likely depends upon voice memory, rather than word memory. There is also evidence to suggest that voice effects can be found in implicit memory paradigms. However, the presence of voice effects appears to depend greatly on the task employed. Using a word identification task, perceptual similarity between study and test conditions is, like for explicit memory tests, crucial. In addition, the type of noise employed appears to have a differential effect. While voice effects have been observed when white noise is used at both study and test, using multi-talker babble does not confer the same results. In terms of neuroimaging research modulations, characterization of an implicit memory effect reflective of voice congruency is currently lacking. Copyright © 2014 Elsevier B.V. All rights reserved.

  10. Women's dietary diversity in rural Bangladesh: Pathways through women's empowerment.

    PubMed

    Sinharoy, Sheela S; Waid, Jillian L; Haardörfer, Regine; Wendt, Amanda; Gabrysch, Sabine; Yount, Kathryn M

    2018-01-01

    The relationship between women's empowerment and women's nutrition is understudied. We aimed to elucidate this relationship by quantifying possible pathways between empowerment and dietary diversity among women in rural Bangladesh. In 2015, we conducted a cross-sectional survey of 2,599 married women ages 15-40 (median: 25) living in 96 settlements of Habiganj District, Bangladesh, as a baseline for the Food and Agricultural Approaches to Reducing Malnutrition trial. We collected data on women's empowerment (highest completed grade of schooling and agency), dietary diversity, and demographic factors, including household wealth. We used exploratory factor analysis and confirmatory factor analysis on random split-half samples, followed by structural equation modelling, to test pathways from schooling, through domains of women's agency, to dietary diversity. Factor analysis revealed 3 latent domains of women's agency: social solidarity, decision-making, and voice with husband. In the adjusted mediation model, having any postprimary schooling was positively associated with voice with husband (β 41  = .051, p = .010), which was positively associated with dietary diversity (β 54  = .39, p = .002). Schooling also had a direct positive association with women's dietary diversity (β 51  = .22, p < .001). Neither women's social solidarity nor decision-making mediated the relationship between schooling and dietary diversity. The link between schooling and dietary diversity was direct and indirect, through women's voice with husband but not through women's social solidarity or decision-making. In this population, women with postprimary schooling seem to be better able to negotiate improved diets for themselves. © 2017 John Wiley & Sons Ltd.

  11. Matching novel face and voice identity using static and dynamic facial images.

    PubMed

    Smith, Harriet M J; Dunn, Andrew K; Baguley, Thom; Stacey, Paula C

    2016-04-01

    Research investigating whether faces and voices share common source identity information has offered contradictory results. Accurate face-voice matching is consistently above chance when the facial stimuli are dynamic, but not when the facial stimuli are static. We tested whether procedural differences might help to account for the previous inconsistencies. In Experiment 1, participants completed a sequential two-alternative forced choice matching task. They either heard a voice and then saw two faces or saw a face and then heard two voices. Face-voice matching was above chance when the facial stimuli were dynamic and articulating, but not when they were static. In Experiment 2, we tested whether matching was more accurate when faces and voices were presented simultaneously. The participants saw two face-voice combinations, presented one after the other. They had to decide which combination was the same identity. As in Experiment 1, only dynamic face-voice matching was above chance. In Experiment 3, participants heard a voice and then saw two static faces presented simultaneously. With this procedure, static face-voice matching was above chance. The overall results, analyzed using multilevel modeling, showed that voices and dynamic articulating faces, as well as voices and static faces, share concordant source identity information. It seems, therefore, that above-chance static face-voice matching is sensitive to the experimental procedure employed. In addition, the inconsistencies in previous research might depend on the specific stimulus sets used; our multilevel modeling analyses show that some people look and sound more similar than others.

  12. [An across-scales analysis of the voice self-concept questionnaire (FESS)].

    PubMed

    Nusseck, Manfred; Richter, Bernhard; Echternach, Matthias; Spahn, Claudia

    2018-04-01

    The questionnaire for the assessment of the voice selfconcept (FESS) contains three sub-scales indicating the personal relation with the own voice. The scales address the relationship with one's own voice, the awareness of the use of one's own voice, and the perception of the connection between voice and emotional changes. A comprehensive approach across the three scales supporting a simplified interpretation of the results was still missing. The FESS questionnaire was used in a sample of 536 German teachers. With a discrimination analysis, commonalities in the scale characteristics were investigated. For a comparative validation with voice health and psychological and physiological wellbeing, the Voice Handicap Index (VHI), the questionnaire for Work-related Behavior and Experience Patterns (AVEM), and the questionnaire for Health-related Quality of Life (SF-12) were additionally collected. The analysis provided four different groups of voice self-concept: group 1 with healthy values in the voice self-concept and wellbeing scales, group 2 with a low voice self-concept and mean wellbeing values, group 3 with a high awareness of the voice use and mean wellbeing values and group 4 with low values in all scales. The results show that a combined approach across all scales of the questionnaire for the assessment of the voice self-concept enables a more detailed interpretation of the characteristics in the voice self-concept. The presented groups provide an applicable use supporting medical diagnoses. © Georg Thieme Verlag KG Stuttgart · New York.

  13. Compact waveguide power divider with multiple isolated outputs

    DOEpatents

    Moeller, Charles P.

    1987-01-01

    A waveguide power divider (10) for splitting electromagnetic microwave power and directionally coupling the divided power includes an input waveguide (21) and reduced height output waveguides (23) interconnected by axial slots (22) and matched loads (25) and (26) positioned at the unused ends of input and output guides (21) and (23) respectively. The axial slots are of a length such that the wave in the input waveguide (21) is directionally coupled to the output waveguides (23). The widths of input guide (21) and output guides (23) are equal and the width of axial slots (22) is one half of the width of the input guide (21).

  14. A Cross-Lingual Mobile Medical Communication System Prototype for Foreigners and Subjects with Speech, Hearing, and Mental Disabilities Based on Pictograms

    PubMed Central

    Wołk, Agnieszka; Glinkowski, Wojciech

    2017-01-01

    People with speech, hearing, or mental impairment require special communication assistance, especially for medical purposes. Automatic solutions for speech recognition and voice synthesis from text are poor fits for communication in the medical domain because they are dependent on error-prone statistical models. Systems dependent on manual text input are insufficient. Recently introduced systems for automatic sign language recognition are dependent on statistical models as well as on image and gesture quality. Such systems remain in early development and are based mostly on minimal hand gestures unsuitable for medical purposes. Furthermore, solutions that rely on the Internet cannot be used after disasters that require humanitarian aid. We propose a high-speed, intuitive, Internet-free, voice-free, and text-free tool suited for emergency medical communication. Our solution is a pictogram-based application that provides easy communication for individuals who have speech or hearing impairment or mental health issues that impair communication, as well as foreigners who do not speak the local language. It provides support and clarification in communication by using intuitive icons and interactive symbols that are easy to use on a mobile device. Such pictogram-based communication can be quite effective and ultimately make people's lives happier, easier, and safer. PMID:29230254

  15. Effects of a Semioccluded Vocal Tract on Laryngeal Muscle Activity and Glottal Adduction in a Single Female Subject

    PubMed Central

    Laukkanen, Anne-Maria; Titze, Ingo R.; Hoffman, Henry; Finnegan, Eileen

    2015-01-01

    Voice training exploits semiocclusives, which increase vocal tract interaction with the source. Modeling results suggest that vocal economy (maximum flow declination rate divided by maximum area declination rate, MADR) is improved by matching the glottal and vocal tract impedances. Changes in MADR may be correlated with thyroarytenoid (TA) muscle activity. Here the effects of impedance matching are studied for laryngeal muscle activity and glottal resistance. One female repeated [pa:p:a] before and immediately after (a) phonation into different-sized tubes and (b) voiced bilabial fricative [β:]. To allow estimation of subglottic pressure from the oral pressure, [p] was inserted also in the repetitions of the semiocclusions. Airflow was registered using a flow mask. EMG was registered from TA, cricothyroid (CT) and lateral cricoarytenoid (LCA) muscles. Phonation was simulated using a 7 × 5 × 5 point-mass model of the vocal folds, allowing inputs of simulated laryngeal muscle activation. The variables were TA, CT and LCA activities. Increased vocal tract impedance caused the subject to raise TA activity compared to CT and LCA activities. Computer simulation showed that higher glottal economy and efficiency (oral radiated power divided by aerodynamic power) were obtained with a higher TA/CT ratio when LCA activity was tuned for ideal adduction. PMID:19011306

  16. A Cross-Lingual Mobile Medical Communication System Prototype for Foreigners and Subjects with Speech, Hearing, and Mental Disabilities Based on Pictograms.

    PubMed

    Wołk, Krzysztof; Wołk, Agnieszka; Glinkowski, Wojciech

    2017-01-01

    People with speech, hearing, or mental impairment require special communication assistance, especially for medical purposes. Automatic solutions for speech recognition and voice synthesis from text are poor fits for communication in the medical domain because they are dependent on error-prone statistical models. Systems dependent on manual text input are insufficient. Recently introduced systems for automatic sign language recognition are dependent on statistical models as well as on image and gesture quality. Such systems remain in early development and are based mostly on minimal hand gestures unsuitable for medical purposes. Furthermore, solutions that rely on the Internet cannot be used after disasters that require humanitarian aid. We propose a high-speed, intuitive, Internet-free, voice-free, and text-free tool suited for emergency medical communication. Our solution is a pictogram-based application that provides easy communication for individuals who have speech or hearing impairment or mental health issues that impair communication, as well as foreigners who do not speak the local language. It provides support and clarification in communication by using intuitive icons and interactive symbols that are easy to use on a mobile device. Such pictogram-based communication can be quite effective and ultimately make people's lives happier, easier, and safer.

  17. Voice symptoms and voice-related quality of life in college students.

    PubMed

    Merrill, Ray M; Tanner, Kristine; Merrill, Joseph G; McCord, Matthew D; Beardsley, Melissa M; Steele, Brittanie A

    2013-08-01

    The purpose of this study was to examine the prevalence of voice disorders in college students and their effect on the students as shown by quality-of-life indicators. A cross-sectional survey was completed by 545 college students in 2012. The survey included 10 questions from the Voice-Related Quality of Life (V-RQOL), selected voice symptoms, and quality-of-life indicators of functional health and well-being based on the Short Form 36-item Health Survey (SF-36). Twenty-nine percent of the college students (mean age, 22.7 years) reported a history of a voice disorder. Hoarseness was the most prevalent voice symptom, but was not correlated with V-RQOL scores. A wobbly or shaky voice, throat dryness, vocal fatigue, and vocal effort explained a significant amount of variance on the social-emotional and physical domains of the V-RQOL index (p < 0.05). Voice symptoms limited emotional and physical functioning as indicated by SF-36 scores. Voice disorders significantly influence psychosocial and physical functioning in college students. These findings have important implications for voice-care services in this population.

  18. Connections between voice ergonomic risk factors in classrooms and teachers' voice production.

    PubMed

    Rantala, Leena M; Hakala, Suvi; Holmqvist, Sofia; Sala, Eeva

    2012-01-01

    The aim of the study was to investigate if voice ergonomic risk factors in classrooms correlated with acoustic parameters of teachers' voice production. The voice ergonomic risk factors in the fields of working culture, working postures and indoor air quality were assessed in 40 classrooms using the Voice Ergonomic Assessment in Work Environment - Handbook and Checklist. Teachers (32 females, 8 males) from the above-mentioned classrooms recorded text readings before and after a working day. Fundamental frequency, sound pressure level (SPL) and the slope of the spectrum (alpha ratio) were analyzed. The higher the number of the risk factors in the classrooms, the higher SPL the teachers used and the more strained the males' voices (increased alpha ratio) were. The SPL was already higher before the working day in the teachers with higher risk than in those with lower risk. In the working environment with many voice ergonomic risk factors, speakers increase voice loudness and use more strained voice quality (males). A practical implication of the results is that voice ergonomic assessments are needed in schools. Copyright © 2013 S. Karger AG, Basel.

  19. In Search of Voice: Theory and Methods in K-12 Student Voice Research in the Us, 1990-2010

    ERIC Educational Resources Information Center

    Gonzalez, Taucia E.; Hernandez-Saca, David I.; Artiles, Alfredo J.

    2017-01-01

    Student voice research is a promising field of study that disrupts traditional student roles by reorganizing learning spaces that center youth voices. This review synthesizes student voice research by answering the following questions: (a) To what extent has student voice been studied at the K-12 levels in the US? (b) What are the conceptual…

  20. Observations of the directional distribution of the wind energy input function over swell waves

    NASA Astrophysics Data System (ADS)

    Shabani, Behnam; Babanin, Alex V.; Baldock, Tom E.

    2016-02-01

    Field measurements of wind stress over shallow water swell traveling in different directions relative to the wind are presented. The directional distribution of the measured stresses is used to confirm the previously proposed but unverified directional distribution of the wind energy input function. The observed wind energy input function is found to follow a much narrower distribution (β∝cos⁡3.6θ) than the Plant (1982) cosine distribution. The observation of negative stress angles at large wind-wave angles, however, indicates that the onset of negative wind shearing occurs at about θ≈ 50°, and supports the use of the Snyder et al. (1981) directional distribution. Taking into account the reverse momentum transfer from swell to the wind, Snyder's proposed parameterization is found to perform exceptionally well in explaining the observed narrow directional distribution of the wind energy input function, and predicting the wind drag coefficients. The empirical coefficient (ɛ) in Snyder's parameterization is hypothesised to be a function of the wave shape parameter, with ɛ value increasing as the wave shape changes between sinusoidal, sawtooth, and sharp-crested shoaling waves.

  1. Varieties of Voice-Hearing: Psychics and the Psychosis Continuum

    PubMed Central

    Powers, Albert R.; Kelley, Megan S.; Corlett, Philip R.

    2017-01-01

    Hearing voices that are not present is a prominent symptom of serious mental illness. However, these experiences may be common in the non-help-seeking population, leading some to propose the existence of a continuum of psychosis from health to disease. Thus far, research on this continuum has focused on what is impaired in help-seeking groups. Here we focus on protective factors in non-help-seeking voice-hearers. We introduce a new study population: clairaudient psychics who receive daily auditory messages. We conducted phenomenological interviews with these subjects, as well as with patients diagnosed with a psychotic disorder who hear voices, people with a diagnosis of a psychotic disorder who do not hear voices, and matched control subjects (without voices or a diagnosis). We found the hallucinatory experiences of psychic voice-hearers to be very similar to those of patients who were diagnosed. We employed techniques from forensic psychiatry to conclude that the psychics were not malingering. Critically, we found that this sample of non-help-seeking voice hearers were able to control the onset and offset of their voices, that they were less distressed by their voice-hearing experiences and that, the first time they admitted to voice-hearing, the reception by others was much more likely to be positive. Patients had much more negative voice-hearing experiences, were more likely to receive a negative reaction when sharing their voices with others for the first time, and this was subsequently more disruptive to their social relationships. We predict that this sub-population of healthy voice-hearers may have much to teach us about the neurobiology, cognitive psychology and ultimately the treatment of voices that are distressing. PMID:28053132

  2. Postlingual adult performance in noise with HiRes 120 and ClearVoice Low, Medium, and High.

    PubMed

    Holden, Laura K; Brenner, Christine; Reeder, Ruth M; Firszt, Jill B

    2013-11-01

    The study's objectives were to evaluate speech recognition in multiple listening conditions using several noise types with HiRes 120 and ClearVoice (Low, Medium, High) and to determine which ClearVoice program was most beneficial for everyday use. Fifteen postlingual adults attended four sessions; speech recognition was assessed at sessions 1 and 3 with HiRes 120 and at sessions 2 and 4 with all ClearVoice programs. Test measures included sentences presented in restaurant noise (R-SPACE), in speech-spectrum noise, in four- and eight-talker babble, and connected discourse presented in 12-talker babble. Participants completed a questionnaire comparing ClearVoice programs. Significant group differences in performance between HiRes 120 and ClearVoice were present only in the R-SPACE; performance was better with ClearVoice High than HiRes 120. Among ClearVoice programs, no significant group differences were present for any measure. Individual results revealed most participants performed better in the R-SPACE with ClearVoice than HiRes 120. For other measures, significant individual differences between HiRes 120 and ClearVoice were not prevalent. Individual results among ClearVoice programs differed and overall preferences varied. Questionnaire data indicated increased understanding with High and Medium in certain environments. R-SPACE and questionnaire results indicated an advantage for ClearVoice High and Medium. Individual test and preference data showed mixed results between ClearVoice programs making global recommendations difficult; however, results suggest providing ClearVoice High and Medium and HiRes 120 as processor options for adults willing to change settings. For adults unwilling or unable to change settings, ClearVoice Medium is a practical choice for daily listening.

  3. The Role of Occupational Voice Demand and Patient-Rated Impairment in Predicting Voice Therapy Adherence.

    PubMed

    Ebersole, Barbara; Soni, Resha S; Moran, Kathleen; Lango, Miriam; Devarajan, Karthik; Jamal, Nausheen

    2018-05-01

    Examine the relationship among the severity of patient-perceived voice impairment, perceptual dysphonia severity, occupational voice demand, and voice therapy adherence. Identify clinical predictors of increased risk for therapy nonadherence. A retrospective cohort study of patients presenting with a chief complaint of persistent dysphonia at an interdisciplinary voice center was done. The Voice Handicap Index-10 (VHI-10) and the Voice-Related Quality of Life (V-RQOL) survey scores, clinician rating of dysphonia severity using the Grade score from the Grade, Roughness Breathiness, Asthenia, and Strain scale, occupational voice demand, and patient demographics were tested for associations with therapy adherence, defined as completion of the treatment plan. Classification and Regression Tree (CART) analysis was performed to establish thresholds for nonadherence risk. Of 166 patients evaluated, 111 were recommended for voice therapy. The therapy nonadherence rate was 56%. Occupational voice demand category, VHI-10, and V-RQOL scores were the only factors significantly correlated with therapy adherence (P < 0.0001, P = 0.018, and P = 0.008, respectively). CART analysis found that patients with low or no occupational voice demand are significantly more likely to be nonadherent with therapy than those with high occupational voice demand (P < 0.001). Furthermore, a VHI-10 score of ≤29 or a V-RQOL score of >40 is a significant cutoff point for predicting therapy nonadherence (P < 0.011 and P < 0.004, respectively). Occupational voice demand and patient perception of impairment are significantly and independently correlated with therapy adherence. A VHI-10 score of ≤9 or a V-RQOL score of >40 is a significant cutoff point for predicting nonadherence risk. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  4. Relationship between Voice Complaints and Subjective and Objective Measures of Vocal Function in Iranian Female Teachers.

    PubMed

    Faham, Maryam; Jalilevand, Nahid; Torabinezhad, Farhad; Silverman, Erin Pearson; Ahmadi, Akram; Anaraki, Zahra Ghayoumi; Jafari, Narges

    2017-07-01

    Teachers are at high risk of developing voice problems because of the excessive vocal demands necessitated by their profession. Teachers' self-assessment of vocal complaints, combined with subjective and objective measures of voice, may enable better therapeutic decision-making. This investigation compared audio-perceptual assessment and acoustic variables in teachers with and without voice complaints. Ninety-nine teachers completed this cross-sectional study and were assigned to one of two groups: those "with voice complaint (VC)" and those "without voice complaint (W-VC)." Voice samples were collected during reading, counting, and vowel prolongation tasks. Teachers were also asked to document any voice symptoms they experienced. Voice samples were analyzed using Dr. Speech program (4th version; Tiger Ltd., USA), and labeled "normal" or "abnormal" according to the "grade" dimension "G" from GRBAS scale. Twenty-one teachers were assigned to the VC group based on self-assessment data. There were statistically significant differences between the two groups with regard to self-reported voice symptoms of hoarseness, breathiness, pitch breaks, and vocal fatigue (P < 0.05). Fourteen participants in the VC group and 40 from the W-VC group were determined to demonstrate "abnormal" vocal quality on perceptual assessment. Only harmonic-to-noise ratio was significantly higher for the W-VC group (ES = 0.55). Teachers with and without voice complaints differed in the incidence, but not type of voice symptoms. Teachers' voice complaints did not correspond to perceptual and acoustic measures. This suggests a potential unmet need for teachers to receive further education on voice disorders. Copyright © 2017 The Voice Foundation. All rights reserved.

  5. The Influence of Sleep Disorders on Voice Quality.

    PubMed

    Rocha, Bruna Rainho; Behlau, Mara

    2017-09-19

    To verify the influence of sleep quality on the voice. Descriptive and analytical cross-sectional study. Data were collected by an online or printed survey divided in three parts: (1) demographic data and vocal health aspects; (2) self-assessment of sleep and vocal quality, and the influence that sleep has on voice; and (3) sleep and voice self-assessment inventories-the Epworth Sleepiness Scale (ESS), the Pittsburgh Sleep Quality Index (PSQI), and the Voice Handicap Index reduced version (VHI-10). A total of 862 people were included (493 women, 369 men), with a mean age of 32 years old (maximum age of 79 and minimum age of 18 years old). The perception of the influence that sleep has on voice showed a difference (P < 0.050) between measures of sleep quality and vocal self-assessment. There were higher scores on the ESS, PSQI, and VHI-10 protocols if sleep and vocal self-assessment were poor. The results indicate that the greater the effect that sleep has on voice, the greater the perceived voice handicap. The aspects that influence a voice handicap are vocal self-assessment, ESS total score, and self-assessment of the influence that sleep has on voice. The absence of daytime sleepiness is a protective factor (odds ratio [OR] > 1) against perceived voice handicap; the presence of daytime sleepiness is a damaging factor (OR < 1). Sleep quality influences voice. Perceived poor sleep quality is related to perceived poor vocal quality. Individuals with a voice handicap observe a greater influence of sleep on voice than those without. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  6. Combined Use of Standard and Throat Microphones for Measurement of Acoustic Voice Parameters and Voice Categorization.

    PubMed

    Uloza, Virgilijus; Padervinskis, Evaldas; Uloziene, Ingrida; Saferis, Viktoras; Verikas, Antanas

    2015-09-01

    The aim of the present study was to evaluate the reliability of the measurements of acoustic voice parameters obtained simultaneously using oral and contact (throat) microphones and to investigate utility of combined use of these microphones for voice categorization. Voice samples of sustained vowel /a/ obtained from 157 subjects (105 healthy and 52 pathological voices) were recorded in a soundproof booth simultaneously through two microphones: oral AKG Perception 220 microphone (AKG Acoustics, Vienna, Austria) and contact (throat) Triumph PC microphone (Clearer Communications, Inc, Burnaby, Canada) placed on the lamina of thyroid cartilage. Acoustic voice signal data were measured for fundamental frequency, percent of jitter and shimmer, normalized noise energy, signal-to-noise ratio, and harmonic-to-noise ratio using Dr. Speech software (Tiger Electronics, Seattle, WA). The correlations of acoustic voice parameters in vocal performance were statistically significant and strong (r = 0.71-1.0) for the entire functional measurements obtained for the two microphones. When classifying into healthy-pathological voice classes, the oral-shimmer revealed the correct classification rate (CCR) of 75.2% and the throat-jitter revealed CCR of 70.7%. However, combination of both throat and oral microphones allowed identifying a set of three voice parameters: throat-signal-to-noise ratio, oral-shimmer, and oral-normalized noise energy, which provided the CCR of 80.3%. The measurements of acoustic voice parameters using a combination of oral and throat microphones showed to be reliable in clinical settings and demonstrated high CCRs when distinguishing the healthy and pathological voice patient groups. Our study validates the suitability of the throat microphone signal for the task of automatic voice analysis for the purpose of voice screening. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  7. Voice parameters and videonasolaryngoscopy in children with vocal nodules: a longitudinal study, before and after voice therapy.

    PubMed

    Valadez, Victor; Ysunza, Antonio; Ocharan-Hernandez, Esther; Garrido-Bustamante, Norma; Sanchez-Valerio, Araceli; Pamplona, Ma C

    2012-09-01

    Vocal Nodules (VN) are a functional voice disorder associated with voice misuse and abuse in children. There are few reports addressing vocal parameters in children with VN, especially after a period of vocal rehabilitation. The purpose of this study is to describe measurements of vocal parameters including Fundamental Frequency (FF), Shimmer (S), and Jitter (J), videonasolaryngoscopy examination and clinical perceptual assessment, before and after voice therapy in children with VN. Voice therapy was provided using visual support through Speech-Viewer software. Twenty patients with VN were studied. An acoustical analysis of voice was performed and compared with data from subjects from a control group matched by age and gender. Also, clinical perceptual assessment of voice and videonasolaryngoscopy were performed to all patients with VN. After a period of voice therapy, provided with visual support using Speech Viewer-III (SV-III-IBM) software, new acoustical analyses, perceptual assessments and videonasolaryngoscopies were performed. Before the onset of voice therapy, there was a significant difference (p<0.05) in mean FF, S and J, between the patients with VN and subjects from the control group. After the voice therapy period, a significant improvement (p<0.05) was found in all acoustic voice parameters. Moreover, perceptual voice analysis demonstrated improvement in all cases. Finally, videonasolaryngoscopy demonstrated that vocal nodules were no longer discernible on the vocal folds in any of the cases. SV-III software seems to be a safe and reliable method for providing voice therapy in children with VN. Acoustic voice parameters, perceptual data and videonasolaryngoscopy were significantly improved after the speech therapy period was completed. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.

  8. Listening to the student voice to improve educational software

    PubMed Central

    van Wyk, Mari; van Ryneveld, Linda

    2017-01-01

    ABSTRACT Academics often develop software for teaching and learning purposes with the best of intentions, only to be disappointed by the low acceptance rate of the software by their students once it is implemented. In this study, the focus is on software that was designed to enable veterinary students to record their clinical skills. A pilot of the software clearly showed that the program had not been received as well as had been anticipated, and therefore the researchers used a group interview and a questionnaire with closed-ended and open-ended questions to obtain the students’ feedback. The open-ended questions were analysed with conceptual content analysis, and themes were identified. Students made valuable suggestions about what they regarded as important considerations when a new software program is introduced. The most important lesson learnt was that students cannot always predict their needs accurately if they are asked for input prior to the development of software. For that reason student input should be obtained on a continuous and regular basis throughout the design and development phases. PMID:28678678

  9. Temporal Stability of Receptiveness to Clinical Research on Alzheimer’s disease

    PubMed Central

    Lingler, Jennifer Hagerty; Rubin, Daniel; Saxton, Judith A.

    2011-01-01

    Research advance directives are a proposed mechanism for ensuring that decisions regarding research participation adhere to preferences voiced by persons with Alzheimer’s disease (AD) prior to losing decisional capacity. While this approach rests on the assumption that preferences regarding research participation are consistent over time, little is known about the stability of such preferences. The purpose of this study was to evaluate the temporal stability of older adults’ receptiveness to participation in clinical trials, neuroimaging studies, and psychosocial investigations on AD. One hundred and four participants in the University of Pittsburgh Alzheimer Disease Research Center (ADRC) were annually surveyed regarding their willingness to be contacted regarding clinical drug trials, neuroimaging studies, and psychosocial research for which they might be eligible. Receptiveness to contact regarding AD research was compared at two time points, one year apart. At baseline, most respondents were willing to be contacted regarding their eligibility for drug trials, imaging studies, and psychosocial research. Thirty-seven percent of respondents voiced a different set of preferences at Year 2 as compared to Year 1. Differences included both increased and decreased willingness to be contacted. Neither stability of preferences nor direction of change (more vs. less willing) varied by diagnostic group. Bivariate analyses revealed that participation in at least one ancillary research study was associated with an overall increase in willingness to be contacted. We conclude that a significant proportion of research-friendly individuals voice different sets of preferences regarding the possibility of research participation when queried at different points in time. Amenability to participating in clinical research on AD is a relatively dynamic personal attribute that may be influenced by personal experience with research participation. This finding has relevance for the policy debate around research advance directives, an approach which assumes that preferences regarding research participation are consistent over time. PMID:20711058

  10. Voice quality after endoscopic laser surgery and radiotherapy for early glottic cancer: objective measurements emphasizing the Voice Handicap Index

    PubMed Central

    Caminero Cueva, Maria Jesús; Señaris González, Blanca; Llorente Pendás, José Luis; Gorriz Gil, Carmen; López Llames, Aurora; Alonso Pantiga, Ramón; Suárez Nieto, Carlos

    2007-01-01

    We analyzed the functional outcome and self-evaluation of the voice of patients with T1 glottic carcinoma treated with endoscopic laser surgery and radiotherapy. We performed an objective voice evaluation, as well as a physical, emotional and functional well being assessment of 19 patients treated with laser surgery and 18 patients treated with radiotherapy. Voice quality is affected both by surgery and radiotherapy. Voice parameters only show differences in the maximum phonation time between both treatments. Results in the Voice Handicap Index show that radiotherapy has less effect on patient voice quality perception. There is a reduced impact on the patient’s perception of voice quality after radiotherapy, despite there being no significant differences in vocal quality between radiotherapy and laser cordectomy. PMID:17999074

  11. Learning [Voice

    ERIC Educational Resources Information Center

    Tauberer, Joshua Ian

    2010-01-01

    The [voice] distinction between homorganic stops and fricatives is made by a number of acoustic correlates including voicing, segment duration, and preceding vowel duration. The present work looks at [voice] from a number of multidimensional perspectives. This dissertation's focus is a corpus study of the phonetic realization of [voice] in two…

  12. Swinging at a cocktail party: voice familiarity aids speech perception in the presence of a competing voice.

    PubMed

    Johnsrude, Ingrid S; Mackey, Allison; Hakyemez, Hélène; Alexander, Elizabeth; Trang, Heather P; Carlyon, Robert P

    2013-10-01

    People often have to listen to someone speak in the presence of competing voices. Much is known about the acoustic cues used to overcome this challenge, but almost nothing is known about the utility of cues derived from experience with particular voices--cues that may be particularly important for older people and others with impaired hearing. Here, we use a version of the coordinate-response-measure procedure to show that people can exploit knowledge of a highly familiar voice (their spouse's) not only to track it better in the presence of an interfering stranger's voice, but also, crucially, to ignore it so as to comprehend a stranger's voice more effectively. Although performance declines with increasing age when the target voice is novel, there is no decline when the target voice belongs to the listener's spouse. This finding indicates that older listeners can exploit their familiarity with a speaker's voice to mitigate the effects of sensory and cognitive decline.

  13. Perceptual connections between prepubertal children's voices in their speaking behavior and their singing behavior.

    PubMed

    Rinta, Tiija Elisabet; Welch, Graham F

    2009-11-01

    Traditionally, children's speaking and singing behaviors have been regarded as two separate sets of behaviors. Nevertheless, according to the voice-scientific view, all vocal functioning is interconnected due to the fact that we exploit the same voice and the same physiological mechanisms in generating all vocalization. The intention of the study was to investigate whether prepubertal children's speaking and singing behaviors are connected perceptually. Voice recordings were conducted with 60 10-year-old children. Each child performed a set of speaking and singing tasks in the voice experiments. Each voice sample was analyzed perceptually with a specially designed perceptual voice assessment protocol. The main finding was that the children's vocal functioning and voice quality in their speaking behavior correlated statistically significantly with those in their singing behavior. The findings imply that children's speaking and singing behaviors are perceptually connected through their vocal functioning and voice quality. Thus, it can be argued that children possess one voice that is used for generating their speaking and singing behaviors.

  14. Comparison of speech perception performance between Sprint/Esprit 3G and Freedom processors in children implanted with nucleus cochlear implants.

    PubMed

    Santarelli, Rosamaria; Magnavita, Vincenzo; De Filippi, Roberta; Ventura, Laura; Genovese, Elisabetta; Arslan, Edoardo

    2009-04-01

    To compare speech perception performance in children fitted with previous generation Nucleus sound processor, Sprint or Esprit 3G, and the Freedom, the most recently released system from the Cochlear Corporation that features a larger input dynamic range. Prospective intrasubject comparative study. University Medical Center. Seventeen prelingually deafened children who had received the Nucleus 24 cochlear implant and used the Sprint or Esprit 3G sound processor. Cochlear implantation with Cochlear device. Speech perception was evaluated at baseline (Sprint, n = 11; Esprit 3G, n = 6) and after 1 month's experience with the Freedom sound processor. Identification and recognition of disyllabic words and identification of vowels were performed via recorded voice in quiet (70 dB [A]), in the presence of background noise at various levels of signal-to-noise ratio (+10, +5, 0, -5) and at a soft presentation level (60 dB [A]). Consonant identification and recognition of disyllabic words, trisyllabic words, and sentences were evaluated in live voice. Frequency discrimination was measured in a subset of subjects (n = 5) by using an adaptive, 3-interval, 3-alternative, forced-choice procedure. Identification of disyllabic words administered at a soft presentation level showed a significant increase when switching to the Freedom compared with the previously worn processor in children using the Sprint or Esprit 3G. Identification and recognition of disyllabic words in the presence of background noise as well as consonant identification and sentence recognition increased significantly for the Freedom compared with the previously worn device only in children fitted with the Sprint. Frequency discrimination was significantly better when switching to the Freedom compared with the previously worn processor. Serial comparisons revealed that that speech perception performance evaluated in children aged 5 to 15 years was superior with the Freedom than previous generations of Nucleus sound processors. These differences are deemed to ensue from an increased input dynamic range, a feature that offers potentially enhanced phonemic discrimination.

  15. Designing interaction, voice, and inclusion in AAC research.

    PubMed

    Pullin, Graham; Treviranus, Jutta; Patel, Rupal; Higginbotham, Jeff

    2017-09-01

    The ISAAC 2016 Research Symposium included a Design Stream that examined timely issues across augmentative and alternative communication (AAC), framed in terms of designing interaction, designing voice, and designing inclusion. Each is a complex term with multiple meanings; together they represent challenging yet important frontiers of AAC research. The Design Stream was conceived by the four authors, researchers who have been exploring AAC and disability-related design throughout their careers, brought together by a shared conviction that designing for communication implies more than ensuring access to words and utterances. Each of these presenters came to AAC from a different background: interaction design, inclusive design, speech science, and social science. The resulting discussion among 24 symposium participants included controversies about the role of technology, tensions about independence and interdependence, and a provocation about taste. The paper concludes by proposing new directions for AAC research: (a) new interdisciplinary research could combine scientific and design research methods, as distant yet complementary as microanalysis and interaction design, (b) new research tools could seed accessible and engaging contextual research into voice within a social model of disability, and (c) new open research networks could support inclusive, international and interdisciplinary research.

  16. Processing of speech signals for physical and sensory disabilities.

    PubMed Central

    Levitt, H

    1995-01-01

    Assistive technology involving voice communication is used primarily by people who are deaf, hard of hearing, or who have speech and/or language disabilities. It is also used to a lesser extent by people with visual or motor disabilities. A very wide range of devices has been developed for people with hearing loss. These devices can be categorized not only by the modality of stimulation [i.e., auditory, visual, tactile, or direct electrical stimulation of the auditory nerve (auditory-neural)] but also in terms of the degree of speech processing that is used. At least four such categories can be distinguished: assistive devices (a) that are not designed specifically for speech, (b) that take the average characteristics of speech into account, (c) that process articulatory or phonetic characteristics of speech, and (d) that embody some degree of automatic speech recognition. Assistive devices for people with speech and/or language disabilities typically involve some form of speech synthesis or symbol generation for severe forms of language disability. Speech synthesis is also used in text-to-speech systems for sightless persons. Other applications of assistive technology involving voice communication include voice control of wheelchairs and other devices for people with mobility disabilities. Images Fig. 4 PMID:7479816

  17. Laryngeal Aerodynamics in Children with Hearing Impairment versus Age and Height Matched Normal Hearing Peers.

    PubMed

    Das, Barshapriya; Chatterjee, Indranil; Kumar, Suman

    2013-01-01

    Lack of proper auditory feedback in hearing-impaired subjects results in functional voice disorder. It is directly related to discoordination of intrinsic and extrinsic laryngeal muscles and disturbed contraction and relaxation of antagonistic muscles. A total of twenty children in the age range of 5-10 years were considered for the study. They were divided into two groups: normal hearing children and hearing aid user children. Results showed a significant difference in the vital capacity, maximum sustained phonation, and fast adduction abduction rate having equal variance for normal and hearing aid user children, respectively, but no significant difference was found in the peak flow value with being statistically significant. A reduced vital capacity in hearing aid user children suggests a limited use of the lung volume for speech production. It may be inferred from the study that the hearing aid user children have poor vocal proficiency which is reflected in their voice. The use of voicing component in hearing impaired subjects is seen due to improper auditory feedback. It was found that there was a significant difference in the vital capacity, maximum sustained phonation (MSP), and fast adduction abduction rate and no significant difference in the peak flow.

  18. Cause-effect relationship between vocal fold physiology and voice production in a three-dimensional phonation model

    PubMed Central

    Zhang, Zhaoyan

    2016-01-01

    The goal of this study is to better understand the cause-effect relation between vocal fold physiology and the resulting vibration pattern and voice acoustics. Using a three-dimensional continuum model of phonation, the effects of changes in vocal fold stiffness, medial surface thickness in the vertical direction, resting glottal opening, and subglottal pressure on vocal fold vibration and different acoustic measures are investigated. The results show that the medial surface thickness has dominant effects on the vertical phase difference between the upper and lower margins of the medial surface, closed quotient, H1-H2, and higher-order harmonics excitation. The main effects of vocal fold approximation or decreasing resting glottal opening are to lower the phonation threshold pressure, reduce noise production, and increase the fundamental frequency. Increasing subglottal pressure is primarily responsible for vocal intensity increase but also leads to significant increase in noise production and an increased fundamental frequency. Increasing AP stiffness significantly increases the fundamental frequency and slightly reduces noise production. The interaction among vocal fold thickness, stiffness, approximation, and subglottal pressure in the control of F0, vocal intensity, and voice quality is discussed. PMID:27106298

  19. [Vascular lesions of vocal folds--part 1: horizontal vascular lesions].

    PubMed

    Voigt-Zimmermann, S; Arens, C

    2014-12-01

    In recent decades, the endoscopic methods and technologies for laryngeal examination have improved so much that not only epithelial changes, but also vascular changes are recognizable at earlier stages. When comparing newer and older literature, the associated increasingly differentiated descriptions of such visible vascular changes of the vocal folds lead to terminological blurring and shifts of meaning. This complicates the technical-scientific discourse. The aim of the present work is a theoretical and conceptual clarification of early vascular changes of vocal folds. Horizontal changes of benigne vascular diseases, e. g. vessel ectasia, meander, increasing number and branching of vessels, change of direction may develop in to manifest vascular lesions, like varicosis, polyps and in case of ruptures to haemorrhages of vocal folds. These beginning and reversible vascular changes, when early detected and discussed basing on etiological knowledge, may lead to more differentiated prognostic statements and adequate therapeutic decisions, e. g. phonosurgery, functional voice therapy, voice hygiene and voice rest. Vertical vascular changes, like vessel loops, occur primarily in laryngeal papilloma, pre-cancerous and cancerous changes of the vocal folds. Already in small cancerous lesions of the vocal folds the vascular architecture is completely destroyed. © Georg Thieme Verlag KG Stuttgart · New York.

  20. Processing of Speech Signals for Physical and Sensory Disabilities

    NASA Astrophysics Data System (ADS)

    Levitt, Harry

    1995-10-01

    Assistive technology involving voice communication is used primarily by people who are deaf, hard of hearing, or who have speech and/or language disabilities. It is also used to a lesser extent by people with visual or motor disabilities. A very wide range of devices has been developed for people with hearing loss. These devices can be categorized not only by the modality of stimulation [i.e., auditory, visual, tactile, or direct electrical stimulation of the auditory nerve (auditory-neural)] but also in terms of the degree of speech processing that is used. At least four such categories can be distinguished: assistive devices (a) that are not designed specifically for speech, (b) that take the average characteristics of speech into account, (c) that process articulatory or phonetic characteristics of speech, and (d) that embody some degree of automatic speech recognition. Assistive devices for people with speech and/or language disabilities typically involve some form of speech synthesis or symbol generation for severe forms of language disability. Speech synthesis is also used in text-to-speech systems for sightless persons. Other applications of assistive technology involving voice communication include voice control of wheelchairs and other devices for people with mobility disabilities.

  1. Using the critical incident technique to define a minimal data set for requirements elicitation in public health.

    PubMed

    Olvingson, Christina; Hallberg, Niklas; Timpka, Toomas; Greenes, Robert A

    2002-12-18

    The introduction of computer-based information systems (ISs) in public health provides enhanced possibilities for service improvements and hence also for improvement of the population's health. Not least, new communication systems can help in the socialization and integration process needed between the different professions and geographical regions. Therefore, development of ISs that truly support public health practices require that technical, cognitive, and social issues be taken into consideration. A notable problem is to capture 'voices' of all potential users, i.e., the viewpoints of different public health practitioners. Failing to capture these voices will result in inefficient or even useless systems. The aim of this study is to develop a minimal data set for capturing users' voices on problems experienced by public health professionals in their daily work and opinions about how these problems can be solved. The issues of concern thus captured can be used both as the basis for formulating the requirements of ISs for public health professionals and to create an understanding of the use context. Further, the data can help in directing the design to the features most important for the users.

  2. Does religious belief enable positive interpretation of auditory hallucinations?: a comparison of religious voice hearers with and without psychosis.

    PubMed

    Cottam, S; Paul, S N; Doughty, O J; Carpenter, L; Al-Mousawi, A; Karvounis, S; Done, D J

    2011-09-01

    Introduction. Hearing voices occurs in people without psychosis. Why hearing voices is such a key pathological feature of psychosis whilst remaining a manageable experience in nonpsychotic people is yet to be understood. We hypothesised that religious voice hearers would interpret voices in accordance with their beliefs and therefore experience less distress. Methods. Three voice hearing groups, which comprised: 20 mentally healthy Christians, 15 Christian patients with psychosis, and 14 nonreligious patients with psychosis. All completed (1) questionnaires with rating scales measuring the perceptual and emotional aspects of hallucinated voices, and (2) a semistructured interview to explore whether religious belief is used to make sense of the voice hearing experience. Results. The three groups had perceptually similar experiences when hearing the voices. Mentally healthy Christians appeared to assimilate the experience with their religious beliefs (schematic processing) resulting in positive interpretations. Christian patients tended not to assimilate the experience with their religious beliefs, frequently reporting nonreligious interpretations that were predominantly negative. Nearly all participants experienced voices as powerful, but mentally healthy Christians reported the power of voices positively. Conclusion. Religious belief appeared to have a profound, beneficial influence on the mentally healthy Christians' interpretation of hearing voices, but had little or no influence in the case of Christian patients.

  3. Voice disorders in teachers and the general population: effects on work performance, attendance, and future career choices.

    PubMed

    Roy, Nelson; Merrill, Ray M; Thibeault, Susan; Gray, Steven D; Smith, Elaine M

    2004-06-01

    To examine the frequency and adverse effects of voice disorders on job performance and attendance in teachers and the general population, 2,401 participants from Iowa and Utah (n1 = 1,243 teachers and n2 = 1,279 nonteachers) were randomly selected and were interviewed by telephone using a voice disorder questionnaire. Teachers were significantly more likely than nonteachers to have experienced multiple voice symptoms and signs including hoarseness, discomfort, and increased effort while using their voice, tiring or experiencing a change in voice quality after short use, difficulty projecting their voice, trouble speaking or singing softly, and a loss of their singing range (all odds ratios [ORs] p <.05). Furthermore, teachers consistently attributed these voice symptoms to their occupation and were significantly more likely to indicate that their voice limited their ability to perform certain tasks at work, and had reduced activities or interactions as a result. Teachers, as compared with nonteachers, had missed more workdays over the preceding year because of voice problems and were more likely to consider changing occupations because of their voice (all comparisons p <.05). These findings strongly suggest that occupationally related voice dysfunction in teachers can have significant adverse effects on job performance, attendance, and future career choices.

  4. Transmasculine People's Voice Function: A Review of the Currently Available Evidence.

    PubMed

    Azul, David; Nygren, Ulrika; Södersten, Maria; Neuschaefer-Rube, Christiane

    2017-03-01

    This study aims to evaluate the currently available discursive and empirical data relating to those aspects of transmasculine people's vocal situations that are not primarily gender-related, to identify restrictions to voice function that have been observed in this population, and to make suggestions for future voice research and clinical practice. We conducted a comprehensive review of the voice literature. Publications were identified by searching six electronic databases and bibliographies of relevant articles. Twenty-two publications met inclusion criteria. Discourses and empirical data were analyzed for factors and practices that impact on voice function and for indications of voice function-related problems in transmasculine people. The quality of the evidence was appraised. The extent and quality of studies investigating transmasculine people's voice function was found to be limited. There was mixed evidence to suggest that transmasculine people might experience restrictions to a range of domains of voice function, including vocal power, vocal control/stability, glottal function, pitch range/variability, vocal endurance, and voice quality. More research into the different factors and practices affecting transmasculine people's voice function that takes account of a range of parameters of voice function and considers participants' self-evaluations is needed to establish how functional voice production can be best supported in this population. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.

  5. A study on the application of voice interaction in automotive human machine interface experience design

    NASA Astrophysics Data System (ADS)

    Huang, Zhaohui; Huang, Xiemin

    2018-04-01

    This paper, firstly, introduces the application trend of the integration of multi-channel interactions in automotive HMI ((Human Machine Interface) from complex information models faced by existing automotive HMI and describes various interaction modes. By comparing voice interaction and touch screen, gestures and other interaction modes, the potential and feasibility of voice interaction in automotive HMI experience design are concluded. Then, the related theories of voice interaction, identification technologies, human beings' cognitive models of voices and voice design methods are further explored. And the research priority of this paper is proposed, i.e. how to design voice interaction to create more humane task-oriented dialogue scenarios to enhance interactive experiences of automotive HMI. The specific scenarios in driving behaviors suitable for the use of voice interaction are studied and classified, and the usability principles and key elements for automotive HMI voice design are proposed according to the scenario features. Then, through the user participatory usability testing experiment, the dialogue processes of voice interaction in automotive HMI are defined. The logics and grammars in voice interaction are classified according to the experimental results, and the mental models in the interaction processes are analyzed. At last, the voice interaction design method to create the humane task-oriented dialogue scenarios in the driving environment is proposed.

  6. Voice responses to changes in pitch of voice or tone auditory feedback

    NASA Astrophysics Data System (ADS)

    Sivasankar, Mahalakshmi; Bauer, Jay J.; Babu, Tara; Larson, Charles R.

    2005-02-01

    The present study was undertaken to examine if a subject's voice F0 responded not only to perturbations in pitch of voice feedback but also to changes in pitch of a side tone presented congruent with voice feedback. Small magnitude brief duration perturbations in pitch of voice or tone auditory feedback were randomly introduced during sustained vowel phonations. Results demonstrated a higher rate and larger magnitude of voice F0 responses to changes in pitch of the voice compared with a triangular-shaped tone (experiment 1) or a pure tone (experiment 2). However, response latencies did not differ across voice or tone conditions. Data suggest that subjects responded to the change in F0 rather than harmonic frequencies of auditory feedback because voice F0 response prevalence, magnitude, or latency did not statistically differ across triangular-shaped tone or pure-tone feedback. Results indicate the audio-vocal system is sensitive to the change in pitch of a variety of sounds, which may represent a flexible system capable of adapting to changes in the subject's voice. However, lower prevalence and smaller responses to tone pitch-shifted signals suggest that the audio-vocal system may resist changes to the pitch of other environmental sounds when voice feedback is present. .

  7. Intra-oral pressure-based voicing control of electrolaryngeal speech with intra-oral vibrator.

    PubMed

    Takahashi, Hirokazu; Nakao, Masayuki; Kikuchi, Yataro; Kaga, Kimitaka

    2008-07-01

    In normal speech, coordinated activities of intrinsic laryngeal muscles suspend a glottal sound at utterance of voiceless consonants, automatically realizing a voicing control. In electrolaryngeal speech, however, the lack of voicing control is one of the causes of unclear voice, voiceless consonants tending to be misheard as the corresponding voiced consonants. In the present work, we developed an intra-oral vibrator with an intra-oral pressure sensor that detected utterance of voiceless phonemes during the intra-oral electrolaryngeal speech, and demonstrated that an intra-oral pressure-based voicing control could improve the intelligibility of the speech. The test voices were obtained from one electrolaryngeal speaker and one normal speaker. We first investigated on the speech analysis software how a voice onset time (VOT) and first formant (F1) transition of the test consonant-vowel syllables contributed to voiceless/voiced contrasts, and developed an adequate voicing control strategy. We then compared the intelligibility of consonant-vowel syllables among the intra-oral electrolaryngeal speech with and without online voicing control. The increase of intra-oral pressure, typically with a peak ranging from 10 to 50 gf/cm2, could reliably identify utterance of voiceless consonants. The speech analysis and intelligibility test then demonstrated that a short VOT caused the misidentification of the voiced consonants due to a clear F1 transition. Finally, taking these results together, the online voicing control, which suspended the prosthetic tone while the intra-oral pressure exceeded 2.5 gf/cm2 and during the 35 milliseconds that followed, proved efficient to improve the voiceless/voiced contrast.

  8. Obligatory and facultative brain regions for voice-identity recognition

    PubMed Central

    Roswandowitz, Claudia; Kappes, Claudia; Obrig, Hellmuth; von Kriegstein, Katharina

    2018-01-01

    Abstract Recognizing the identity of others by their voice is an important skill for social interactions. To date, it remains controversial which parts of the brain are critical structures for this skill. Based on neuroimaging findings, standard models of person-identity recognition suggest that the right temporal lobe is the hub for voice-identity recognition. Neuropsychological case studies, however, reported selective deficits of voice-identity recognition in patients predominantly with right inferior parietal lobe lesions. Here, our aim was to work towards resolving the discrepancy between neuroimaging studies and neuropsychological case studies to find out which brain structures are critical for voice-identity recognition in humans. We performed a voxel-based lesion-behaviour mapping study in a cohort of patients (n = 58) with unilateral focal brain lesions. The study included a comprehensive behavioural test battery on voice-identity recognition of newly learned (voice-name, voice-face association learning) and familiar voices (famous voice recognition) as well as visual (face-identity recognition) and acoustic control tests (vocal-pitch and vocal-timbre discrimination). The study also comprised clinically established tests (neuropsychological assessment, audiometry) and high-resolution structural brain images. The three key findings were: (i) a strong association between voice-identity recognition performance and right posterior/mid temporal and right inferior parietal lobe lesions; (ii) a selective association between right posterior/mid temporal lobe lesions and voice-identity recognition performance when face-identity recognition performance was factored out; and (iii) an association of right inferior parietal lobe lesions with tasks requiring the association between voices and faces but not voices and names. The results imply that the right posterior/mid temporal lobe is an obligatory structure for voice-identity recognition, while the inferior parietal lobe is only a facultative component of voice-identity recognition in situations where additional face-identity processing is required. PMID:29228111

  9. Obligatory and facultative brain regions for voice-identity recognition.

    PubMed

    Roswandowitz, Claudia; Kappes, Claudia; Obrig, Hellmuth; von Kriegstein, Katharina

    2018-01-01

    Recognizing the identity of others by their voice is an important skill for social interactions. To date, it remains controversial which parts of the brain are critical structures for this skill. Based on neuroimaging findings, standard models of person-identity recognition suggest that the right temporal lobe is the hub for voice-identity recognition. Neuropsychological case studies, however, reported selective deficits of voice-identity recognition in patients predominantly with right inferior parietal lobe lesions. Here, our aim was to work towards resolving the discrepancy between neuroimaging studies and neuropsychological case studies to find out which brain structures are critical for voice-identity recognition in humans. We performed a voxel-based lesion-behaviour mapping study in a cohort of patients (n = 58) with unilateral focal brain lesions. The study included a comprehensive behavioural test battery on voice-identity recognition of newly learned (voice-name, voice-face association learning) and familiar voices (famous voice recognition) as well as visual (face-identity recognition) and acoustic control tests (vocal-pitch and vocal-timbre discrimination). The study also comprised clinically established tests (neuropsychological assessment, audiometry) and high-resolution structural brain images. The three key findings were: (i) a strong association between voice-identity recognition performance and right posterior/mid temporal and right inferior parietal lobe lesions; (ii) a selective association between right posterior/mid temporal lobe lesions and voice-identity recognition performance when face-identity recognition performance was factored out; and (iii) an association of right inferior parietal lobe lesions with tasks requiring the association between voices and faces but not voices and names. The results imply that the right posterior/mid temporal lobe is an obligatory structure for voice-identity recognition, while the inferior parietal lobe is only a facultative component of voice-identity recognition in situations where additional face-identity processing is required. © The Author (2017). Published by Oxford University Press on behalf of the Guarantors of Brain.

  10. Voice to Voice: Developing In-Service Teachers' Personal, Collaborative, and Public Voices.

    ERIC Educational Resources Information Center

    Thurber, Frances; Zimmerman, Enid

    1997-01-01

    Describes a model for inservice education that begins with an interchange of teachers' voices with those of the students in an interactive dialog. The exchange allows them to develop their private voices through self-reflection and validation of their own experiences. (JOW)

  11. Voices on Voice: Perspectives, Definitions, Inquiry.

    ERIC Educational Resources Information Center

    Yancey, Kathleen Blake, Ed.

    This collection of essays approaches "voice" as a means of expression that lives in the interactions of writers, readers, and language, and examines the conceptualizations of voice within the oral rhetorical and expressionist traditions, and the notion of voice as both a singular and plural phenomenon. An explanatory introduction by the…

  12. Voice Therapy Practices and Techniques: A Survey of Voice Clinicians.

    ERIC Educational Resources Information Center

    Mueller, Peter B.; Larson, George W.

    1992-01-01

    Eighty-three voice disorder therapists' ratings of statements regarding voice therapy practices indicated that vocal nodules are the most frequent disorder treated; vocal abuse and hard glottal attack elimination, counseling, and relaxation were preferred treatment approaches; and voice therapy is more effective with adults than with children.…

  13. Predictors of level of voice in adolescent girls: ethnicity, attachment, and gender role socialization.

    PubMed

    Theran, Sally A

    2009-09-01

    The current study empirically examined predictors of level of voice (ethnicity, attachment, and gender role socialization) in a diverse sample of 108 14-year-old girls. Structural equation modeling results indicated that parental attachment predicted level of voice with authority figures, and gender role socialization predicted level of voice with authority figures and peers. Both masculinity and femininity were salient for higher levels of voice with authority figures whereas higher scores on masculinity contributed to higher levels of voice with peers. These findings suggest that, contrary to previous theoretical work, femininity itself is not a risk factor for low levels of voice. In addition, African-American girls had higher levels of voice with teachers and classmates than did Caucasian girls, and girls who were in a school with a greater concentration of ethnic minorities had higher levels of voice with peers than did girls at a school with fewer minority students.

  14. Speech technology and cinema: can they learn from each other?

    PubMed

    Pauletto, Sandra

    2013-10-01

    The voice is the most important sound of a film soundtrack. It represents a character and it carries language. There are different types of cinematic voices: dialogue, internal monologues, and voice-overs. Conventionally, two main characteristics differentiate these voices: lip synchronization and the voice's attributes that make it appropriate for the character (for example, a voice that sounds very close to the audience can be appropriate for a narrator, but not for an onscreen character). What happens, then, if a film character can only speak through an asynchronous machine that produces a 'robot-like' voice? This article discusses the sound-related work and experimentation done by the author for the short film Voice by Choice. It also attempts to discover whether speech technology design can learn from its cinematic representation, and if such uncommon film protagonists can contribute creatively to transform the conventions of cinematic voices.

  15. Dimensionality in voice quality.

    PubMed

    Bele, Irene Velsvik

    2007-05-01

    This study concerns speaking voice quality in a group of male teachers (n = 35) and male actors (n = 36), as the purpose was to investigate normal and supranormal voices. The goal was the development of a method of valid perceptual evaluation for normal to supranormal and resonant voices. The voices (text reading at two loudness levels) had been evaluated by 10 listeners, for 15 vocal characteristics using VA scales. In this investigation, the results of an exploratory factor analysis of the vocal characteristics used in this method are presented, reflecting four dimensions of major importance for normal and supranormal voices. Special emphasis is placed on the effects on voice quality of a change in the loudness variable, as two loudness levels are studied. Furthermore, the vocal characteristics Sonority and Ringing voice quality are paid special attention, as the essence of the term "resonant voice" was a basic issue throughout a doctoral dissertation where this study was included.

  16. When the face fits: recognition of celebrities from matching and mismatching faces and voices.

    PubMed

    Stevenage, Sarah V; Neil, Greg J; Hamlin, Iain

    2014-01-01

    The results of two experiments are presented in which participants engaged in a face-recognition or a voice-recognition task. The stimuli were face-voice pairs in which the face and voice were co-presented and were either "matched" (same person), "related" (two highly associated people), or "mismatched" (two unrelated people). Analysis in both experiments confirmed that accuracy and confidence in face recognition was consistently high regardless of the identity of the accompanying voice. However accuracy of voice recognition was increasingly affected as the relationship between voice and accompanying face declined. Moreover, when considering self-reported confidence in voice recognition, confidence remained high for correct responses despite the proportion of these responses declining across conditions. These results converged with existing evidence indicating the vulnerability of voice recognition as a relatively weak signaller of identity, and results are discussed in the context of a person-recognition framework.

  17. Voice quality change in future professional voice users after 9 months of voice training.

    PubMed

    Timmermans, Bernadette; De Bodt, Marc; Wuyts, Floris; Van de Heyning, Paul

    2004-01-01

    Sixty-eight students of a school for audiovisual communication participated in this study. A part of them, 49 students, received voice training for 9 months (the trained group); 19 subjects received no specific voice training (the untrained group). A multidimensional test battery containing the GRBAS scale, videolaryngostroboscopy, Maximum Phonation Time (MPT), jitter, lowest intensity (IL), highest frequency (FoH), Dysphonia Severity Index (DSI) and Voice Handicap Index (VHI) was applied before and after training to evaluate training outcome. The voice training is made up of technical workshops in small groups (five to eight subjects) and vocal coaching in the ateliers. In the technical workshops, basic skills are trained (posture, breathing technique, articulation and diction), and in the ateliers, the speech and language pathologist assists the subjects in the practice of their voice work. This study revealed a significant amelioration over time for the objective measurements [Dysphonia Severity Index: from 2.3 to 4.5 ( P<0.001)] and the self-evaluation [Voice Handicap Index, from 23 to 18.4 ( P=0.016)] for the trained group only. This outcome favors the systematic introduction of voice training during the schooling of professional voice users.

  18. A suggestion to improve a day keeps your depletion away: Examining promotive and prohibitive voice behaviors within a regulatory focus and ego depletion framework.

    PubMed

    Lin, Szu-Han Joanna; Johnson, Russell E

    2015-09-01

    One way that employees contribute to organizational effectiveness is by expressing voice. They may offer suggestions for how to improve the organization (promotive voice behavior), or express concerns to prevent harmful events from occurring (prohibitive voice behavior). Although promotive and prohibitive voices are thought to be distinct types of behavior, very little is known about their unique antecedents and consequences. In this study we draw on regulatory focus and ego depletion theories to derive a theoretical model that outlines a dynamic process of the antecedents and consequences of voice behavior. Results from 2 multiwave field studies revealed that promotion and prevention foci have unique ties to promotive and prohibitive voice, respectively. Promotive and prohibitive voice, in turn, were associated with decreases and increases, respectively, in depletion. Consistent with the dynamic nature of self-control, depletion was associated with reductions in employees' subsequent voice behavior, regardless of the type of voice (promotive or prohibitive). Results were consistent across 2 studies and remained even after controlling for other established antecedents of voice and alternative mediating mechanisms beside depletion. (c) 2015 APA, all rights reserved).

  19. Acoustic and perceptual characteristics of the voice in patients with vocal polyps after surgery and voice therapy.

    PubMed

    Petrovic-Lazic, Mirjana; Jovanovic, Nadica; Kulic, Milan; Babac, Snezana; Jurisic, Vladimir

    2015-03-01

    The aim of the study was to assess the effect of endolaryngeal phonomicrosurgery (EPM) and voice therapy in patients with vocal fold polyps using perceptual and acoustic analysis before and after both therapies. The acoustic tests and perceptual evaluation of voice were carried out on 41 female patients with vocal fold polyp before and after EPM and voice therapy. Both therapy strategies were performed. Used acoustic parameters were Jitter percent (Jitt), pitch perturbation quotient (PPQ), shimmer percent (Shim), amplitude perturbation quotient (APQ), fundamental frequency variation (vF0), noise-to-harmonic ratio (NHR), Voice Turbulence Index (VTI). For perceptual evaluation, GRB scale was used. Results indicated higher values of investigated parameters in patients' group than in the control group (P < 0.01). Good correlation between the perceptual hoarseness factors of GRB scale and objective acoustic voice parameters were observed. All analyzed acoustic parameters improved after the phonomicrosurgery and voice therapy and tend to approach to values of the control group. For Jitt percent, Shim percent, vF0, VTI, and NHR, there were statistically significant differences. Perceptual voice evaluation revealed statistically significantly (P < 0.01) decreased rating of G (grade), R (rough) and B (breathy) after surgery and voice therapy. Our data indicated that both acoustic and perceptual characteristic of voice in patients with vocal polyps significantly improved after phonomicrosurgical and voice treatment. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  20. Comparison of Voice Handicap Index Scores Between Female Students of Speech Therapy and Other Health Professions.

    PubMed

    Tafiadis, Dionysios; Chronopoulos, Spyridon K; Siafaka, Vassiliki; Drosos, Konstantinos; Kosma, Evangelia I; Toki, Eugenia I; Ziavra, Nausica

    2017-09-01

    Students' groups (eg, teachers, speech language pathologists) are presumably at risk of developing a voice disorder due to misuse of their voice, which will affect their way of living. Multidisciplinary voice assessment of student populations is currently spread widely along with the use of self-reported questionnaires. This study compared the Voice Handicap Index domains and item scores between female students of speech and language therapy and of other health professions in Greece. We also examined the probability of speech language therapy students developing any vocal symptom. Two hundred female non-dysphonic students (aged 18-31) were recruited. Participants answered the Voice Evaluation Form and the Greek adaptation of the Voice Handicap Index. Significant differences were observed between the two groups (students of speech therapy and other health professions) through Voice Handicap Index (total score, functional and physical domains), excluding the emotional domain. Furthermore, significant differences for specific Voice Handicap Index items, between subgroups, were observed. In conclusion, speech language therapy students had higher Voice Handicap Index scores, which probably could be an indicator for avoiding profession-related dysphonia at a later stage. Also, Voice Handicap Index could be at a first glance an assessment tool for the recognition of potential voice disorder development in students. In turn, the results could be used for indirect therapy approaches, such as providing methods for maintaining vocal health in different student populations. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

Top