Sample records for voice response system

  1. 75 FR 30845 - Request Voucher for Grant Payment and Line of Credit Control System (LOCCS) Voice Response System...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-06-02

    ... request vouchers for distribution of grant funds using the automated Voice Response System (VRS). An... Payment and Line of Credit Control System (LOCCS) Voice Response System Access Authorization AGENCY... subject proposal. Payment request vouchers for distribution of grant funds using the automated Voice...

  2. Single-channel voice-response-system program documentation volume I : system description

    DOT National Transportation Integrated Search

    1977-01-01

    This report documents the design and implementation of a Voice Response System (VRS) using Adaptive Differential Pulse Code Modulation (ADPCM) voice coding. Implemented on a Digital Equipment Corporation PDP-11/20,R this VRS system supports a single ...

  3. Voice responses to changes in pitch of voice or tone auditory feedback

    NASA Astrophysics Data System (ADS)

    Sivasankar, Mahalakshmi; Bauer, Jay J.; Babu, Tara; Larson, Charles R.

    2005-02-01

    The present study was undertaken to examine if a subject's voice F0 responded not only to perturbations in pitch of voice feedback but also to changes in pitch of a side tone presented congruent with voice feedback. Small magnitude brief duration perturbations in pitch of voice or tone auditory feedback were randomly introduced during sustained vowel phonations. Results demonstrated a higher rate and larger magnitude of voice F0 responses to changes in pitch of the voice compared with a triangular-shaped tone (experiment 1) or a pure tone (experiment 2). However, response latencies did not differ across voice or tone conditions. Data suggest that subjects responded to the change in F0 rather than harmonic frequencies of auditory feedback because voice F0 response prevalence, magnitude, or latency did not statistically differ across triangular-shaped tone or pure-tone feedback. Results indicate the audio-vocal system is sensitive to the change in pitch of a variety of sounds, which may represent a flexible system capable of adapting to changes in the subject's voice. However, lower prevalence and smaller responses to tone pitch-shifted signals suggest that the audio-vocal system may resist changes to the pitch of other environmental sounds when voice feedback is present. .

  4. Voice Response Systems Technology.

    ERIC Educational Resources Information Center

    Gerald, Jeanette

    1984-01-01

    Examines two methods of generating synthetic speech in voice response systems, which allow computers to communicate in human terms (speech), using human interface devices (ears): phoneme and reconstructed voice systems. Considerations prior to implementation, current and potential applications, glossary, directory, and introduction to Input Output…

  5. The Voice as Computer Interface: A Look at Tomorrow's Technologies.

    ERIC Educational Resources Information Center

    Lange, Holley R.

    1991-01-01

    Discussion of voice as the communications device for computer-human interaction focuses on voice recognition systems for use within a library environment. Voice technologies are described, including voice response and voice recognition; examples of voice systems in use in libraries are examined; and further possibilities, including use with…

  6. 78 FR 71676 - Submission for Review: 3206-0201, Federal Employees Health Benefits (FEHB) Open Season Express...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-11-29

    ... (FEHB) Open Season Express Interactive Voice Response (IVR) System and Open Season Web site AGENCY: U.S... Benefits (FEHB) Open Season Express Interactive Voice Response (IVR) System and the Open Season Web site... Season Express Interactive Voice Response (IVR) System, and the Open Season Web site, Open Season Online...

  7. Twenty-Channel Voice Response System

    DOT National Transportation Integrated Search

    1981-06-01

    This report documents the design and implementation of a Voice Response System, which provides Direct-User Access to the FAA's aviation-weather data base. This system supports 20 independent audio channels, and as of this report, speaks three weather...

  8. Vocal responses to unanticipated perturbations in voice loudness feedback: an automatic mechanism for stabilizing voice amplitude.

    PubMed

    Bauer, Jay J; Mittal, Jay; Larson, Charles R; Hain, Timothy C

    2006-04-01

    The present study tested whether subjects respond to unanticipated short perturbations in voice loudness feedback with compensatory responses in voice amplitude. The role of stimulus magnitude (+/- 1,3 vs 6 dB SPL), stimulus direction (up vs down), and the ongoing voice amplitude level (normal vs soft) were compared across compensations. Subjects responded to perturbations in voice loudness feedback with a compensatory change in voice amplitude 76% of the time. Mean latency of amplitude compensation was 157 ms. Mean response magnitudes were smallest for 1-dB stimulus perturbations (0.75 dB) and greatest for 6-dB conditions (0.98 dB). However, expressed as gain, responses for 1-dB perturbations were largest and almost approached 1.0. Response magnitudes were larger for the soft voice amplitude condition compared to the normal voice amplitude condition. A mathematical model of the audio-vocal system captured the main features of the compensations. Previous research has demonstrated that subjects can respond to an unanticipated perturbation in voice pitch feedback with an automatic compensatory response in voice fundamental frequency. Data from the present study suggest that voice loudness feedback can be used in a similar manner to monitor and stabilize voice amplitude around a desired loudness level.

  9. Voice Response System Statistics Program : Operational Handbook.

    DOT National Transportation Integrated Search

    1980-06-01

    This report documents the Voice Response System (VRS) Statistics Program developed for the preflight weather briefing VRS. It describes the VRS statistical report format and contents, the software program structure, and the program operation.

  10. National Voice Response System (VRS) Implementation Plan Alternatives Study

    DOT National Transportation Integrated Search

    1979-07-01

    This study examines the alternatives available to implement a national Voice Response System (VRS) for automated preflight weather briefings and flight plan filing. Four major hardware configurations are discussed. A computerized analysis model was d...

  11. Response time effects of alerting tone and semantic context for synthesized voice cockpit warnings

    NASA Technical Reports Server (NTRS)

    Simpson, C. A.; Williams, D. H.

    1980-01-01

    Some handbooks and human factors design guides have recommended that a voice warning should be preceded by a tone to attract attention to the warning. As far as can be determined from a search of the literature, no experimental evidence supporting this exists. A fixed-base simulator flown by airline pilots was used to test the hypothesis that the total 'system-time' to respond to a synthesized voice cockpit warning would be longer when the message was preceded by a tone because the voice itself was expected to perform both the alerting and the information transfer functions. The simulation included realistic ATC radio voice communications, synthesized engine noise, cockpit conversation, and realistic flight routes. The effect of a tone before a voice warning was to lengthen response time; that is, responses were slower with an alerting tone. Lengthening the voice warning with another work, however, did not increase response time.

  12. 78 FR 31972 - Notice of Proposed Information Collection for Public Comment; Request Voucher for Grant Payment...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-05-28

    ... request vouchers for distribution of grant funds using the automated Voice Response System (VRS). An... Information Collection for Public Comment; Request Voucher for Grant Payment and Line of Credit Control System (LOCCS) Voice Response System Access AGENCY: Office of the Chief Financial Officer, HUD. ACTION: Notice...

  13. Effects of Voice Harmonic Complexity on ERP Responses to Pitch-Shifted Auditory Feedback

    PubMed Central

    Behroozmand, Roozbeh; Korzyukov, Oleg; Larson, Charles R.

    2011-01-01

    Objective The present study investigated the neural mechanisms of voice pitch control for different levels of harmonic complexity in the auditory feedback. Methods Event-related potentials (ERPs) were recorded in response to +200 cents pitch perturbations in the auditory feedback of self-produced natural human vocalizations, complex and pure tone stimuli during active vocalization and passive listening conditions. Results During active vocal production, ERP amplitudes were largest in response to pitch shifts in the natural voice, moderately large for non-voice complex stimuli and smallest for the pure tones. However, during passive listening, neural responses were equally large for pitch shifts in voice and non-voice complex stimuli but still larger than that for pure tones. Conclusions These findings suggest that pitch change detection is facilitated for spectrally rich sounds such as natural human voice and non-voice complex stimuli compared with pure tones. Vocalization-induced increase in neural responses for voice feedback suggests that sensory processing of naturally-produced complex sounds such as human voice is enhanced by means of motor-driven mechanisms (e.g. efference copies) during vocal production. Significance This enhancement may enable the audio-vocal system to more effectively detect and correct for vocal errors in the feedback of natural human vocalizations to maintain an intended vocal output for speaking. PMID:21719346

  14. 76 FR 72306 - Federal Housing Administration (FHA) Appraiser Roster: Appraiser Qualifications for Placement on...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-11-23

    ... Appraiser Roster regulations by replacing the obsolete references to the Credit Alert Interactive Voice Response System (CAIVRS) with references to its successor, the online-based Credit Alert Verification... propose the elimination references to the Credit Alert Interactive Voice Response System (CAIVRS). On July...

  15. 76 FR 41441 - Federal Housing Administration (FHA) Appraiser Roster: Appraiser Qualifications for Placement on...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-07-14

    ... the FHA Appraiser Roster by replacing the obsolete references to the Credit Alert Interactive Voice Response System with references to its successor, the online-based Credit Alert Verification Reporting...'s Limited Denial of Participation list, or in HUD's Credit Alert Interactive Voice Response System...

  16. Effects of voice harmonic complexity on ERP responses to pitch-shifted auditory feedback.

    PubMed

    Behroozmand, Roozbeh; Korzyukov, Oleg; Larson, Charles R

    2011-12-01

    The present study investigated the neural mechanisms of voice pitch control for different levels of harmonic complexity in the auditory feedback. Event-related potentials (ERPs) were recorded in response to+200 cents pitch perturbations in the auditory feedback of self-produced natural human vocalizations, complex and pure tone stimuli during active vocalization and passive listening conditions. During active vocal production, ERP amplitudes were largest in response to pitch shifts in the natural voice, moderately large for non-voice complex stimuli and smallest for the pure tones. However, during passive listening, neural responses were equally large for pitch shifts in voice and non-voice complex stimuli but still larger than that for pure tones. These findings suggest that pitch change detection is facilitated for spectrally rich sounds such as natural human voice and non-voice complex stimuli compared with pure tones. Vocalization-induced increase in neural responses for voice feedback suggests that sensory processing of naturally-produced complex sounds such as human voice is enhanced by means of motor-driven mechanisms (e.g. efference copies) during vocal production. This enhancement may enable the audio-vocal system to more effectively detect and correct for vocal errors in the feedback of natural human vocalizations to maintain an intended vocal output for speaking. Copyright © 2011 International Federation of Clinical Neurophysiology. Published by Elsevier Ireland Ltd. All rights reserved.

  17. Sequoyah Foreign Language Translation System - Business Case Analysis

    DTIC Science & Technology

    2007-12-01

    Interactive Natural Dialogue System (S-MINDS)..................................................................20 j. Voice Response Translator ( VRT ...20 Figure 8. U.S. Marine Military Policeman Demonstrating VRT (From: Ref. U.S...www.languagerealm.com/Files/usmc_mt_test_2004.pdf. 21 j. Voice Response Translator ( VRT ) The VRT is a S2S human language translation device that uses

  18. Audio-vocal system regulation in children with autism spectrum disorders.

    PubMed

    Russo, Nicole; Larson, Charles; Kraus, Nina

    2008-06-01

    Do children with autism spectrum disorders (ASD) respond similarly to perturbations in auditory feedback as typically developing (TD) children? Presentation of pitch-shifted voice auditory feedback to vocalizing participants reveals a close coupling between the processing of auditory feedback and vocal motor control. This paradigm was used to test the hypothesis that abnormalities in the audio-vocal system would negatively impact ASD compensatory responses to perturbed auditory feedback. Voice fundamental frequency (F(0)) was measured while children produced an /a/ sound into a microphone. The voice signal was fed back to the subjects in real time through headphones. During production, the feedback was pitch shifted (-100 cents, 200 ms) at random intervals for 80 trials. Averaged voice F(0) responses to pitch-shifted stimuli were calculated and correlated with both mental and language abilities as tested via standardized tests. A subset of children with ASD produced larger responses to perturbed auditory feedback than TD children, while the other children with ASD produced significantly lower response magnitudes. Furthermore, robust relationships between language ability, response magnitude and time of peak magnitude were identified. Because auditory feedback helps to stabilize voice F(0) (a major acoustic cue of prosody) and individuals with ASD have problems with prosody, this study identified potential mechanisms of dysfunction in the audio-vocal system for voice pitch regulation in some children with ASD. Objectively quantifying this deficit may inform both the assessment of a subgroup of ASD children with prosody deficits, as well as remediation strategies that incorporate pitch training.

  19. Synthesized speech rate and pitch effects on intelligibility of warning messages for pilots

    NASA Technical Reports Server (NTRS)

    Simpson, C. A.; Marchionda-Frost, K.

    1984-01-01

    In civilian and military operations, a future threat-warning system with a voice display could warn pilots of other traffic, obstacles in the flight path, and/or terrain during low-altitude helicopter flights. The present study was conducted to learn whether speech rate and voice pitch of phoneme-synthesized speech affects pilot accuracy and response time to typical threat-warning messages. Helicopter pilots engaged in an attention-demanding flying task and listened for voice threat warnings presented in a background of simulated helicopter cockpit noise. Performance was measured by flying-task performance, threat-warning intelligibility, and response time. Pilot ratings were elicited for the different voice pitches and speech rates. Significant effects were obtained only for response time and for pilot ratings, both as a function of speech rate. For the few cases when pilots forgot to respond to a voice message, they remembered 90 percent of the messages accurately when queried for their response 8 to 10 sec later.

  20. Error-dependent modulation of speech-induced auditory suppression for pitch-shifted voice feedback.

    PubMed

    Behroozmand, Roozbeh; Larson, Charles R

    2011-06-06

    The motor-driven predictions about expected sensory feedback (efference copies) have been proposed to play an important role in recognition of sensory consequences of self-produced motor actions. In the auditory system, this effect was suggested to result in suppression of sensory neural responses to self-produced voices that are predicted by the efference copies during vocal production in comparison with passive listening to the playback of the identical self-vocalizations. In the present study, event-related potentials (ERPs) were recorded in response to upward pitch shift stimuli (PSS) with five different magnitudes (0, +50, +100, +200 and +400 cents) at voice onset during active vocal production and passive listening to the playback. Results indicated that the suppression of the N1 component during vocal production was largest for unaltered voice feedback (PSS: 0 cents), became smaller as the magnitude of PSS increased to 200 cents, and was almost completely eliminated in response to 400 cents stimuli. Findings of the present study suggest that the brain utilizes the motor predictions (efference copies) to determine the source of incoming stimuli and maximally suppresses the auditory responses to unaltered feedback of self-vocalizations. The reduction of suppression for 50, 100 and 200 cents and its elimination for 400 cents pitch-shifted voice auditory feedback support the idea that motor-driven suppression of voice feedback leads to distinctly different sensory neural processing of self vs. non-self vocalizations. This characteristic may enable the audio-vocal system to more effectively detect and correct for unexpected errors in the feedback of self-produced voice pitch compared with externally-generated sounds.

  1. Interactive Voice/Web Response System in clinical research

    PubMed Central

    Ruikar, Vrishabhsagar

    2016-01-01

    Emerging technologies in computer and telecommunication industry has eased the access to computer through telephone. An Interactive Voice/Web Response System (IxRS) is one of the user friendly systems for end users, with complex and tailored programs at its backend. The backend programs are specially tailored for easy understanding of users. Clinical research industry has experienced revolution in methodologies of data capture with time. Different systems have evolved toward emerging modern technologies and tools in couple of decades from past, for example, Electronic Data Capture, IxRS, electronic patient reported outcomes, etc. PMID:26952178

  2. Interactive Voice/Web Response System in clinical research.

    PubMed

    Ruikar, Vrishabhsagar

    2016-01-01

    Emerging technologies in computer and telecommunication industry has eased the access to computer through telephone. An Interactive Voice/Web Response System (IxRS) is one of the user friendly systems for end users, with complex and tailored programs at its backend. The backend programs are specially tailored for easy understanding of users. Clinical research industry has experienced revolution in methodologies of data capture with time. Different systems have evolved toward emerging modern technologies and tools in couple of decades from past, for example, Electronic Data Capture, IxRS, electronic patient reported outcomes, etc.

  3. Error-dependent modulation of speech-induced auditory suppression for pitch-shifted voice feedback

    PubMed Central

    2011-01-01

    Background The motor-driven predictions about expected sensory feedback (efference copies) have been proposed to play an important role in recognition of sensory consequences of self-produced motor actions. In the auditory system, this effect was suggested to result in suppression of sensory neural responses to self-produced voices that are predicted by the efference copies during vocal production in comparison with passive listening to the playback of the identical self-vocalizations. In the present study, event-related potentials (ERPs) were recorded in response to upward pitch shift stimuli (PSS) with five different magnitudes (0, +50, +100, +200 and +400 cents) at voice onset during active vocal production and passive listening to the playback. Results Results indicated that the suppression of the N1 component during vocal production was largest for unaltered voice feedback (PSS: 0 cents), became smaller as the magnitude of PSS increased to 200 cents, and was almost completely eliminated in response to 400 cents stimuli. Conclusions Findings of the present study suggest that the brain utilizes the motor predictions (efference copies) to determine the source of incoming stimuli and maximally suppresses the auditory responses to unaltered feedback of self-vocalizations. The reduction of suppression for 50, 100 and 200 cents and its elimination for 400 cents pitch-shifted voice auditory feedback support the idea that motor-driven suppression of voice feedback leads to distinctly different sensory neural processing of self vs. non-self vocalizations. This characteristic may enable the audio-vocal system to more effectively detect and correct for unexpected errors in the feedback of self-produced voice pitch compared with externally-generated sounds. PMID:21645406

  4. Towards a Metalanguage Adequate to Linguistic Achievement in Post-Structuralism and English: Reflections on Voicing in the Writing of Secondary Students

    ERIC Educational Resources Information Center

    Macken-Horarik, Mary; Morgan, Wendy

    2011-01-01

    This paper considers the development of voicing in the writing of secondary English students influenced by post-structuralist approaches to literature. It investigates students' growing capacity not only to voice their own responses to literature but also to relate these to a range of theoretical discourses. Drawing on systemic functional…

  5. Neural effects of environmental advertising: An fMRI analysis of voice age and temporal framing.

    PubMed

    Casado-Aranda, Luis-Alberto; Martínez-Fiestas, Myriam; Sánchez-Fernández, Juan

    2018-01-15

    Ecological information offered to society through advertising enhances awareness of environmental issues, encourages development of sustainable attitudes and intentions, and can even alter behavior. This paper, by means of functional Magnetic Resonance Imaging (fMRI) and self-reports, explores the underlying mechanisms of processing ecological messages. The study specifically examines brain and behavioral responses to persuasive ecological messages that differ in temporal framing and in the age of the voice pronouncing them. The findings reveal that attitudes are more positive toward future-framed messages presented by young voices. The whole-brain analysis reveals that future-framed (FF) ecological messages trigger activation in brain areas related to imagery, prospective memories and episodic events, thus reflecting the involvement of past behaviors in future ecological actions. Past-framed messages (PF), in turn, elicit brain activations within the episodic system. Young voices (YV), in addition to triggering stronger activation in areas involved with the processing of high-timbre, high-pitched and high-intensity voices, are perceived as more emotional and motivational than old voices (OV) as activations in anterior cingulate cortex and amygdala. Messages expressed by older voices, in turn, exhibit stronger activation in areas formerly linked to low-pitched voices and voice gender perception. Interestingly, a link is identified between neural and self-report responses indicating that certain brain activations in response to future-framed messages and young voices predicted higher attitudes toward future-framed and young voice advertisements, respectively. The results of this study provide invaluable insight into the unconscious origin of attitudes toward environmental messages and indicate which voice and temporal frame of a message generate the greatest subconscious value. Copyright © 2017 Elsevier Ltd. All rights reserved.

  6. Voice and choice in health care in England: understanding citizen responses to dissatisfaction.

    PubMed

    Dowding, Keith; John, Peter

    2011-01-01

    Using data from a five-year online survey the paper examines the effects of relative satisfaction with health services on individuals' voice-and-choice activity in the English public health care system. Voice is considered in three parts – individual voice (complaints), collective voice voting and participation (collective action). Exercising choice is seen in terms of complete exit (not using health care), internal exit (choosing another public service provider) and private exit (using private health care). The interaction of satisfaction and forms of voice and choice are analysed over time. Both voice and choice are correlated with dissatisfaction with those who are unhappy with the NHS more likely to privately voice and to plan to take up private health care. Those unable to choose private provision are likely to use private voice. These factors are not affected by items associated with social capital – indeed, being more trusting leads to lower voice activity.

  7. Internet-Based System for Voice Communication With the ISS

    NASA Technical Reports Server (NTRS)

    Chamberlain, James; Myers, Gerry; Clem, David; Speir, Terri

    2005-01-01

    The Internet Voice Distribution System (IVoDS) is a voice-communication system that comprises mainly computer hardware and software. The IVoDS was developed to supplement and eventually replace the Enhanced Voice Distribution System (EVoDS), which, heretofore, has constituted the terrestrial subsystem of a system for voice communications among crewmembers of the International Space Station (ISS), workers at the Payloads Operations Center at Marshall Space Flight Center, principal investigators at diverse locations who are responsible for specific payloads, and others. The IVoDS utilizes a communication infrastructure of NASA and NASArelated intranets in addition to, as its name suggests, the Internet. Whereas the EVoDS utilizes traditional circuitswitched telephony, the IVoDS is a packet-data system that utilizes a voice over Internet protocol (VOIP). Relative to the EVoDS, the IVoDS offers advantages of greater flexibility and lower cost for expansion and reconfiguration. The IVoDS is an extended version of a commercial Internet-based voice conferencing system that enables each user to participate in only one conference at a time. In the IVoDS, a user can receive audio from as many as eight conferences simultaneously while sending audio to one of them. The IVoDS also incorporates administrative controls, beyond those of the commercial system, that provide greater security and control of the capabilities and authorizations for talking and listening afforded to each user.

  8. Brain systems mediating voice identity processing in blind humans.

    PubMed

    Hölig, Cordula; Föcker, Julia; Best, Anna; Röder, Brigitte; Büchel, Christian

    2014-09-01

    Blind people rely more on vocal cues when they recognize a person's identity than sighted people. Indeed, a number of studies have reported better voice recognition skills in blind than in sighted adults. The present functional magnetic resonance imaging study investigated changes in the functional organization of neural systems involved in voice identity processing following congenital blindness. A group of congenitally blind individuals and matched sighted control participants were tested in a priming paradigm, in which two voice stimuli (S1, S2) were subsequently presented. The prime (S1) and the target (S2) were either from the same speaker (person-congruent voices) or from two different speakers (person-incongruent voices). Participants had to classify the S2 as either a old or a young person. Person-incongruent voices (S2) compared with person-congruent voices elicited an increased activation in the right anterior fusiform gyrus in congenitally blind individuals but not in matched sighted control participants. In contrast, only matched sighted controls showed a higher activation in response to person-incongruent compared with person-congruent voices (S2) in the right posterior superior temporal sulcus. These results provide evidence for crossmodal plastic changes of the person identification system in the brain after visual deprivation. Copyright © 2014 Wiley Periodicals, Inc.

  9. The use of an automated interactive voice response system to manage medication identification calls to a poison center.

    PubMed

    Krenzelok, Edward P; Mrvos, Rita

    2009-05-01

    In 2007, medication identification requests (MIRs) accounted for 26.2% of all calls to U.S. poison centers. MIRs are documented with minimal information, but they still require an inordinate amount of work by specialists in poison information (SPI). An analysis was undertaken to identify options to reduce the impact of MIRs on both human and financial resources. All MIRs (2003-2007) to a certified regional poison information center were analyzed to determine call patterns and staffing. The data were used to justify an efficient and cost-effective solution. MIRs represented 42.3% of the 2007 call volume. Optimal staffing would require hiring an additional four full-time equivalent SPI. An interactive voice response (IVR) system was developed to respond to the MIRs. The IVR was used to develop the Medication Identification System that allowed the diversion of up to 50% of the MIRs, enhancing surge capacity and allowing specialists to address the more emergent poison exposure calls. This technology is an entirely voice-activated response call management system that collects zip code, age, gender and drug data and stores all responses as .csv files for reporting purposes. The query bank includes the 200 most common MIRs, and the system features text-to-voice synthesis that allows easy modification of the drug identification menu. Callers always have the option of engaging a SPI at any time during the IVR call flow. The IVR is an efficient and effective alternative that creates better staff utilization.

  10. Vocalization-Induced Enhancement of the Auditory Cortex Responsiveness during Voice F0 Feedback Perturbation

    PubMed Central

    Behroozmand, Roozbeh; Karvelis, Laura; Liu, Hanjun; Larson, Charles R.

    2009-01-01

    Objective The present study investigated whether self-vocalization enhances auditory neural responsiveness to voice pitch feedback perturbation and how this vocalization-induced neural modulation can be affected by the extent of the feedback deviation. Method Event related potentials (ERPs) were recorded in 15 subjects in response to +100, +200 and +500 cents pitch-shifted voice auditory feedback during active vocalization and passive listening to the playback of the self-produced vocalizations. Result The amplitude of the evoked P1 (latency: 73.51 ms) and P2 (latency: 199.55 ms) ERP components in response to feedback perturbation were significantly larger during vocalization than listening. The difference between P2 peak amplitudes during vocalization vs. listening was shown to be significantly larger for +100 than +500 cents stimulus. Conclusion Results indicate that the human auditory cortex is more responsive to voice F0 feedback perturbations during vocalization than passive listening. Greater vocalization-induced enhancement of the auditory responsiveness to smaller feedback perturbations may imply that the audio-vocal system detects and corrects for errors in vocal production that closely match the expected vocal output. Significance Findings of this study support previous suggestions regarding the enhanced auditory sensitivity to feedback alterations during self-vocalization, which may serve the purpose of feedback-based monitoring of one’s voice. PMID:19520602

  11. Explaining the high voice superiority effect in polyphonic music: evidence from cortical evoked potentials and peripheral auditory models.

    PubMed

    Trainor, Laurel J; Marie, Céline; Bruce, Ian C; Bidelman, Gavin M

    2014-02-01

    Natural auditory environments contain multiple simultaneously-sounding objects and the auditory system must parse the incoming complex sound wave they collectively create into parts that represent each of these individual objects. Music often similarly requires processing of more than one voice or stream at the same time, and behavioral studies demonstrate that human listeners show a systematic perceptual bias in processing the highest voice in multi-voiced music. Here, we review studies utilizing event-related brain potentials (ERPs), which support the notions that (1) separate memory traces are formed for two simultaneous voices (even without conscious awareness) in auditory cortex and (2) adults show more robust encoding (i.e., larger ERP responses) to deviant pitches in the higher than in the lower voice, indicating better encoding of the former. Furthermore, infants also show this high-voice superiority effect, suggesting that the perceptual dominance observed across studies might result from neurophysiological characteristics of the peripheral auditory system. Although musically untrained adults show smaller responses in general than musically trained adults, both groups similarly show a more robust cortical representation of the higher than of the lower voice. Finally, years of experience playing a bass-range instrument reduces but does not reverse the high voice superiority effect, indicating that although it can be modified, it is not highly neuroplastic. Results of new modeling experiments examined the possibility that characteristics of middle-ear filtering and cochlear dynamics (e.g., suppression) reflected in auditory nerve firing patterns might account for the higher-voice superiority effect. Simulations show that both place and temporal AN coding schemes well-predict a high-voice superiority across a wide range of interval spacings and registers. Collectively, we infer an innate, peripheral origin for the higher-voice superiority observed in human ERP and psychophysical music listening studies. Copyright © 2013 Elsevier B.V. All rights reserved.

  12. Compensation for pitch-shifted auditory feedback during the production of Mandarin tone sequences

    NASA Astrophysics Data System (ADS)

    Xu, Yi; Larson, Charles R.; Bauer, Jay J.; Hain, Timothy C.

    2004-08-01

    Recent research has found that while speaking, subjects react to perturbations in pitch of voice auditory feedback by changing their voice fundamental frequency (F0) to compensate for the perceived pitch-shift. The long response latencies (150-200 ms) suggest they may be too slow to assist in on-line control of the local pitch contour patterns associated with lexical tones on a syllable-to-syllable basis. In the present study, we introduced pitch-shifted auditory feedback to native speakers of Mandarin Chinese while they produced disyllabic sequences /ma ma/ with different tonal combinations at a natural speaking rate. Voice F0 response latencies (100-150 ms) to the pitch perturbations were shorter than syllable durations reported elsewhere. Response magnitudes increased from 50 cents during static tone to 85 cents during dynamic tone productions. Response latencies and peak times decreased in phrases involving a dynamic change in F0. The larger response magnitudes and shorter latency and peak times in tasks requiring accurate, dynamic control of F0, indicate this automatic system for regulation of voice F0 may be task-dependent. These findings suggest that auditory feedback may be used to help regulate voice F0 during production of bi-tonal Mandarin phrases.

  13. Plastic reorganization of neural systems for perception of others in the congenitally blind.

    PubMed

    Fairhall, S L; Porter, K B; Bellucci, C; Mazzetti, M; Cipolli, C; Gobbini, M I

    2017-09-01

    Recent evidence suggests that the function of the core system for face perception might extend beyond visual face-perception to a broader role in person perception. To critically test the broader role of core face-system in person perception, we examined the role of the core system during the perception of others in 7 congenitally blind individuals and 15 sighted subjects by measuring their neural responses using fMRI while they listened to voices and performed identity and emotion recognition tasks. We hypothesised that in people who have had no visual experience of faces, core face-system areas may assume a role in the perception of others via voices. Results showed that emotions conveyed by voices can be decoded in homologues of the core face system only in the blind. Moreover, there was a specific enhancement of response to verbal as compared to non-verbal stimuli in bilateral fusiform face areas and the right posterior superior temporal sulcus showing that the core system also assumes some language-related functions in the blind. These results indicate that, in individuals with no history of visual experience, areas of the core system for face perception may assume a role in aspects of voice perception that are relevant to social cognition and perception of others' emotions. Copyright © 2017 The Author(s). Published by Elsevier Inc. All rights reserved.

  14. Designing of Intelligent Multilingual Patient Reported Outcome System (IMPROS)

    PubMed Central

    Pourasghar, Faramarz; Partovi, Yeganeh

    2015-01-01

    Background: By self-reporting outcome procedure the patients themselves record disease symptoms outside medical centers and then report them to medical staff in specific periods of time. One of the self-reporting methods is the application of interactive voice response (IVR), in which some pre-designed questions in the form of voice tracks would be played and then the caller responses the questions by pressing phone’s keypad bottoms. Aim: The present research explains the main framework of such system designing according to IVR technology that is for the first time designed and administered in Iran. Methods: Interactive Voice Response system was composed by two main parts of hardware and software. Hardware section includes one or several digital phone lines, a modem card with voice playing capability and a PC. IVR software on the other hand, acts as an intelligent control center, records call information and controls incoming data. Results: One of the main features of the system is its capability to be administered in common PCs, utilizing simple and cheap modems, high speed to take responses and it’s appropriateness to low literate patients. The system is applicable for monitoring chronic diseases, cancer and also in psychological diseases and can be suitable for taking care of elders and Children who require long term cares. Other features include user-friendly, decrease in direct and indirect costs of disease treatment and enjoying from high level of security to access patients’ profiles. Conclusions: Intelligent multilingual patient reported outcome system (IMPROS) by controlling diseases gives the opportunity to patients to have more participation during treatment and it improves mutual interaction between patient and medical staff. Moreover it increases the quality of medical services, Additional to empowering patients and their followers. PMID:26635441

  15. Mobility, Aspiration, Voice: A New Structure of Feeling for Student Equity in Higher Education

    ERIC Educational Resources Information Center

    Sellar, Sam; Gale, Trevor

    2011-01-01

    There is a changed "structure of feeling" emerging in higher education systems, particularly in OECD nations, in response to changed social, cultural and economic arrangements. Taking a student equity perspective, the paper names this change in terms of "mobility", "aspiration" and "voice". It argues that…

  16. Infant face interest is associated with voice information and maternal psychological health.

    PubMed

    Taylor, Gemma; Slade, Pauline; Herbert, Jane S

    2014-11-01

    Early infant interest in their mother's face is driven by an experience based face processing system, and is associated with maternal psychological health, even within a non clinical community sample. The present study examined the role of the voice in eliciting infants' interest in mother and stranger faces and in the association between infant face interest and maternal psychological health. Infants aged 3.5-months were shown photographs of their mother's and a stranger's face paired with an audio recording of their mother's and a stranger's voice that was either matched (e.g., mother's face and voice) or mismatched (e.g., mother's face and stranger's voice). Infants spent more time attending to the stranger's matched face and voice than the mother's matched face and voice and the mismatched faces and voices. Thus, infants demonstrated an earlier preference for a stranger's face when given voice information than when the face is presented alone. In the present sample, maternal psychological health varied with 56.7% of mothers reporting mild mood symptoms (depression, anxiety or stress response to childbirth). Infants of mothers with significant mild maternal mood symptoms looked longer at the faces and voices compared to infants of mothers who did not report mild maternal mood symptoms. In sum, infants' experience based face processing system is sensitive to their mothers' maternal psychological health and the multimodal nature of faces. Copyright © 2014 Elsevier Inc. All rights reserved.

  17. Accuracy and Speed of Response to Different Voice Types in a Cockpit Voice Warning System

    DTIC Science & Technology

    1983-09-01

    military aircraft. Different levels of engine background noise, signal to noise ratio of the warning message, and precursor delivery formats were used. The...flight deck signals, the Society of Automotive Engineers stated that a unique, attention-getting sound (such as a chime, 4 etc.) together with voice...aircr,-ft wherein there is no flight engineer position" (cited in Thorburn, 1971, p. 3). The AFIAS letter cited several incidents in which the VWS had

  18. A 4.8 kbps code-excited linear predictive coder

    NASA Technical Reports Server (NTRS)

    Tremain, Thomas E.; Campbell, Joseph P., Jr.; Welch, Vanoy C.

    1988-01-01

    A secure voice system STU-3 capable of providing end-to-end secure voice communications (1984) was developed. The terminal for the new system will be built around the standard LPC-10 voice processor algorithm. The performance of the present STU-3 processor is considered to be good, its response to nonspeech sounds such as whistles, coughs and impulse-like noises may not be completely acceptable. Speech in noisy environments also causes problems with the LPC-10 voice algorithm. In addition, there is always a demand for something better. It is hoped that LPC-10's 2.4 kbps voice performance will be complemented with a very high quality speech coder operating at a higher data rate. This new coder is one of a number of candidate algorithms being considered for an upgraded version of the STU-3 in late 1989. The problems of designing a code-excited linear predictive (CELP) coder to provide very high quality speech at a 4.8 kbps data rate that can be implemented on today's hardware are considered.

  19. Temporal voice areas exist in autism spectrum disorder but are dysfunctional for voice identity recognition

    PubMed Central

    Borowiak, Kamila; von Kriegstein, Katharina

    2016-01-01

    The ability to recognise the identity of others is a key requirement for successful communication. Brain regions that respond selectively to voices exist in humans from early infancy on. Currently, it is unclear whether dysfunction of these voice-sensitive regions can explain voice identity recognition impairments. Here, we used two independent functional magnetic resonance imaging studies to investigate voice processing in a population that has been reported to have no voice-sensitive regions: autism spectrum disorder (ASD). Our results refute the earlier report that individuals with ASD have no responses in voice-sensitive regions: Passive listening to vocal, compared to non-vocal, sounds elicited typical responses in voice-sensitive regions in the high-functioning ASD group and controls. In contrast, the ASD group had a dysfunction in voice-sensitive regions during voice identity but not speech recognition in the right posterior superior temporal sulcus/gyrus (STS/STG)—a region implicated in processing complex spectrotemporal voice features and unfamiliar voices. The right anterior STS/STG correlated with voice identity recognition performance in controls but not in the ASD group. The findings suggest that right STS/STG dysfunction is critical for explaining voice recognition impairments in high-functioning ASD and show that ASD is not characterised by a general lack of voice-sensitive responses. PMID:27369067

  20. Effect of tonal native language on voice fundamental frequency responses to pitch feedback perturbations during sustained vocalizations

    PubMed Central

    Liu, Hanjun; Wang, Emily Q.; Chen, Zhaocong; Liu, Peng; Larson, Charles R.; Huang, Dongfeng

    2010-01-01

    The purpose of this cross-language study was to examine whether the online control of voice fundamental frequency (F0) during vowel phonation is influenced by language experience. Native speakers of Cantonese and Mandarin, both tonal languages spoken in China, participated in the experiments. Subjects were asked to vocalize a vowel sound ∕u∕ at their comfortable habitual F0, during which their voice pitch was unexpectedly shifted (±50, ±100, ±200, or ±500 cents, 200 ms duration) and fed back instantaneously to them over headphones. The results showed that Cantonese speakers produced significantly smaller responses than Mandarin speakers when the stimulus magnitude varied from 200 to 500 cents. Further, response magnitudes decreased along with the increase in stimulus magnitude in Cantonese speakers, which was not observed in Mandarin speakers. These findings suggest that online control of voice F0 during vocalization is sensitive to language experience. Further, systematic modulations of vocal responses across stimulus magnitude were observed in Cantonese speakers but not in Mandarin speakers, which indicates that this highly automatic feedback mechanism is sensitive to the specific tonal system of each language. PMID:21218905

  1. Real-Time Reconfigurable Adaptive Speech Recognition Command and Control Apparatus and Method

    NASA Technical Reports Server (NTRS)

    Salazar, George A. (Inventor); Haynes, Dena S. (Inventor); Sommers, Marc J. (Inventor)

    1998-01-01

    An adaptive speech recognition and control system and method for controlling various mechanisms and systems in response to spoken instructions and in which spoken commands are effective to direct the system into appropriate memory nodes, and to respective appropriate memory templates corresponding to the voiced command is discussed. Spoken commands from any of a group of operators for which the system is trained may be identified, and voice templates are updated as required in response to changes in pronunciation and voice characteristics over time of any of the operators for which the system is trained. Provisions are made for both near-real-time retraining of the system with respect to individual terms which are determined not be positively identified, and for an overall system training and updating process in which recognition of each command and vocabulary term is checked, and in which the memory templates are retrained if necessary for respective commands or vocabulary terms with respect to an operator currently using the system. In one embodiment, the system includes input circuitry connected to a microphone and including signal processing and control sections for sensing the level of vocabulary recognition over a given period and, if recognition performance falls below a given level, processing audio-derived signals for enhancing recognition performance of the system.

  2. The Belt voice: Acoustical measurements and esthetic correlates

    NASA Astrophysics Data System (ADS)

    Bounous, Barry Urban

    This dissertation explores the esthetic attributes of the Belt voice through spectral acoustical analysis. The process of understanding the nature and safe practice of Belt is just beginning, whereas the understanding of classical singing is well established. The unique nature of the Belt sound provides difficulties for voice teachers attempting to evaluate the quality and appropriateness of a particular sound or performance. This study attempts to provide answers to the question "does Belt conform to a set of measurable esthetic standards?" In answering this question, this paper expands on a previous study of the esthetic attributes of the classical baritone voice (see "Vocal Beauty", NATS Journal 51,1) which also drew some tentative conclusions about the Belt voice but which had an inadequate sample pool of subjects from which to draw. Further, this study demonstrates that it is possible to scientifically investigate the realm of musical esthetics in the singing voice. It is possible to go beyond the "a trained voice compared to an untrained voice" paradigm when evaluating quantitative vocal parameters and actually investigate what truly beautiful voices do. There are functions of sound energy (measured in dB) transference which may affect the nervous system in predictable ways and which can be measured and associated with esthetics. This study does not show consistency in measurements for absolute beauty (taste) even among belt teachers and researchers but does show some markers with varying degrees of importance which may point to a difference between our cognitive learned response to singing and our emotional, more visceral response to sounds. The markers which are significant in determining vocal beauty are: (1) Vibrancy-Characteristics of vibrato including speed, width, and consistency (low variability). (2) Spectral makeup-Ratio of partial strength above the fundamental to the fundamental. (3) Activity of the voice-The quantity of energy being produced. (4) Consistency of the voice-How low is the variability in the energy patterns of the voice.

  3. Vocal Responses to Perturbations in Voice Auditory Feedback in Individuals with Parkinson's Disease

    PubMed Central

    Liu, Hanjun; Wang, Emily Q.; Metman, Leo Verhagen; Larson, Charles R.

    2012-01-01

    Background One of the most common symptoms of speech deficits in individuals with Parkinson's disease (PD) is significantly reduced vocal loudness and pitch range. The present study investigated whether abnormal vocalizations in individuals with PD are related to sensory processing of voice auditory feedback. Perturbations in loudness or pitch of voice auditory feedback are known to elicit short latency, compensatory responses in voice amplitude or fundamental frequency. Methodology/Principal Findings Twelve individuals with Parkinson's disease and 13 age- and sex- matched healthy control subjects sustained a vowel sound (/α/) and received unexpected, brief (200 ms) perturbations in voice loudness (±3 or 6 dB) or pitch (±100 cents) auditory feedback. Results showed that, while all subjects produced compensatory responses in their voice amplitude or fundamental frequency, individuals with PD exhibited larger response magnitudes than the control subjects. Furthermore, for loudness-shifted feedback, upward stimuli resulted in shorter response latencies than downward stimuli in the control subjects but not in individuals with PD. Conclusions/Significance The larger response magnitudes in individuals with PD compared with the control subjects suggest that processing of voice auditory feedback is abnormal in PD. Although the precise mechanisms of the voice feedback processing are unknown, results of this study suggest that abnormal voice control in individuals with PD may be related to dysfunctional mechanisms of error detection or correction in sensory feedback processing. PMID:22448258

  4. Voice disorders and mental health in teachers: a cross-sectional nationwide study.

    PubMed

    Nerrière, Eléna; Vercambre, Marie-Noël; Gilbert, Fabien; Kovess-Masféty, Viviane

    2009-10-02

    Teachers, as professional voice users, are at particular risk of voice disorders. Among contributing factors, stress and psychological tension could play a role but epidemiological data on this problem are scarce. The aim of this study was to evaluate prevalence and cofactors of voice disorders among teachers in the French National Education system, with particular attention paid to the association between voice complaint and psychological status. The source data come from an epidemiological postal survey on physical and mental health conducted in a sample of 20,099 adults (in activity or retired) selected at random from the health plan records of the national education system. Overall response rate was 53%. Of the 10,288 respondents, 3,940 were teachers in activity currently giving classes to students. In the sample of those with complete data (n = 3,646), variables associated with voice disorders were investigated using logistic regression models. Studied variables referred to demographic characteristics, socio-professional environment, psychological distress, mental health disorders (DSM-IV), and sick leave. One in two female teachers reported voice disorders (50.0%) compared to one in four males (26.0%). Those who reported voice disorders presented higher level of psychological distress. Sex- and age-adjusted odds ratios [95% confidence interval] were respectively 1.8 [1.5-2.2] for major depressive episode, 1.7 [1.3-2.2] for general anxiety disorder, and 1.6 [1.2-2.2] for phobia. A significant association between voice disorders and sick leave was also demonstrated (1.5 [1.3-1.7]). Voice disorders were frequent among French teachers. Associations with psychiatric disorders suggest that a situation may exist which is more complex than simple mechanical failure. Further longitudinal research is needed to clarify the comorbidity between voice and psychological disorders.

  5. An effect of loudness of advisory speech on a choice response task

    NASA Astrophysics Data System (ADS)

    Utsuki, Narisuke; Takeuchi, Yoshinori; Nomiyama, Takenori

    1995-03-01

    Recent technologies have realized talking advisory/guidance systems in which machines give advice and guidance to operators in speech. However, nonverbal aspects of spoken messages may have significant effects on an operator's behavior. Twelve subjects participated in a TV game-like choice response task where they were asked to choose a 'true' target from three invader-like figures displayed on a CRT screen. The subjects had received a prerecorded advice designating either left, center, or right target that would be true before each choice. The position of the 'true' targets and advice were preprogrammed in pseudorandom sequences. In other words, there was no way for the subjects to predict the 'true' target and there was no relationship between spoken advice and the true target position. The subjects tended to make more choices corresponding to the presented messages when the messages were presented in a louder voice than in a softer voice. Choice response time was significantly shorter when the response was the same as the advice indicated. The shortening of response time was slightly greater when advice was presented in a louder voice. This study demonstrates that spoken advice may result in faster and less deliberate reponses in accordance with the presented messages which are given by talking guidance systems.

  6. Involvement of the left insula in the ecological validity of the human voice

    PubMed Central

    Tamura, Yuri; Kuriki, Shinji; Nakano, Tamami

    2015-01-01

    A subtle difference between a real human and an artificial object that resembles a human evokes an impression of a large qualitative difference between them. This suggests the existence of a neural mechanism that processes the sense of humanness. To examine the presence of such a mechanism, we compared the behavioral and brain responses of participants who listened to human and artificial singing voices created from vocal fragments of a real human voice. The behavioral experiment showed that the song sung by human voices more often elicited positive feelings and feelings of humanness than the same song sung by artificial voices, although the lyrics, melody, and rhythm were identical. Functional magnetic resonance imaging revealed significantly higher activation in the left posterior insula in response to human voices than in response to artificial voices. Insular activation was not merely evoked by differences in acoustic features between the voices. Therefore, these results suggest that the left insula participates in the neural processing of the ecological quality of the human voice. PMID:25739519

  7. Exploring the anatomical encoding of voice with a mathematical model of the vocal system.

    PubMed

    Assaneo, M Florencia; Sitt, Jacobo; Varoquaux, Gael; Sigman, Mariano; Cohen, Laurent; Trevisan, Marcos A

    2016-11-01

    The faculty of language depends on the interplay between the production and perception of speech sounds. A relevant open question is whether the dimensions that organize voice perception in the brain are acoustical or depend on properties of the vocal system that produced it. One of the main empirical difficulties in answering this question is to generate sounds that vary along a continuum according to the anatomical properties the vocal apparatus that produced them. Here we use a mathematical model that offers the unique possibility of synthesizing vocal sounds by controlling a small set of anatomically based parameters. In a first stage the quality of the synthetic voice was evaluated. Using specific time traces for sub-glottal pressure and tension of the vocal folds, the synthetic voices generated perceptual responses, which are indistinguishable from those of real speech. The synthesizer was then used to investigate how the auditory cortex responds to the perception of voice depending on the anatomy of the vocal apparatus. Our fMRI results show that sounds are perceived as human vocalizations when produced by a vocal system that follows a simple relationship between the size of the vocal folds and the vocal tract. We found that these anatomical parameters encode the perceptual vocal identity (male, female, child) and show that the brain areas that respond to human speech also encode vocal identity. On the basis of these results, we propose that this low-dimensional model of the vocal system is capable of generating realistic voices and represents a novel tool to explore the voice perception with a precise control of the anatomical variables that generate speech. Furthermore, the model provides an explanation of how auditory cortices encode voices in terms of the anatomical parameters of the vocal system. Copyright © 2016 Elsevier Inc. All rights reserved.

  8. It's not what you hear, it's the way you think about it: appraisals as determinants of affect and behaviour in voice hearers.

    PubMed

    Peters, E R; Williams, S L; Cooke, M A; Kuipers, E

    2012-07-01

    Previous studies have suggested that beliefs about voices mediate the relationship between actual voice experience and behavioural and affective response. We investigated beliefs about voice power (omnipotence), voice intent (malevolence/benevolence) and emotional and behavioural response (resistance/engagement) using the Beliefs About Voices Questionnaire - Revised (BAVQ-R) in 46 voice hearers. Distress was assessed using a wide range of measures: voice-related distress, depression, anxiety, self-esteem and suicidal ideation. Voice topography was assessed using measures of voice severity, frequency and intensity. We predicted that beliefs about voices would show a stronger association with distress than voice topography. Omnipotence had the strongest associations with all measures of distress included in the study whereas malevolence was related to resistance, and benevolence to engagement. As predicted, voice severity, frequency and intensity were not related to distress once beliefs were accounted for. These results concur with previous findings that beliefs about voice power are key determinants of distress in voice hearers, and should be targeted specifically in psychological interventions.

  9. Normal voice processing after posterior superior temporal sulcus lesion.

    PubMed

    Jiahui, Guo; Garrido, Lúcia; Liu, Ran R; Susilo, Tirta; Barton, Jason J S; Duchaine, Bradley

    2017-10-01

    The right posterior superior temporal sulcus (pSTS) shows a strong response to voices, but the cognitive processes generating this response are unclear. One possibility is that this activity reflects basic voice processing. However, several fMRI and magnetoencephalography findings suggest instead that pSTS serves as an integrative hub that combines voice and face information. Here we investigate whether right pSTS contributes to basic voice processing by testing Faith, a patient whose right pSTS was resected, with eight behavioral tasks assessing voice identity perception and recognition, voice sex perception, and voice expression perception. Faith performed normally on all the tasks. Her normal performance indicates right pSTS is not necessary for intact voice recognition and suggests that pSTS activations to voices reflect higher-level processes. Copyright © 2017 Elsevier Ltd. All rights reserved.

  10. Voice tracking and spoken word recognition in the presence of other voices

    NASA Astrophysics Data System (ADS)

    Litong-Palima, Marisciel; Violanda, Renante; Saloma, Caesar

    2004-12-01

    We study the human hearing process by modeling the hair cell as a thresholded Hopf bifurcator and compare our calculations with experimental results involving human subjects in two different multi-source listening tasks of voice tracking and spoken-word recognition. In the model, we observed noise suppression by destructive interference between noise sources which weakens the effective noise strength acting on the hair cell. Different success rate characteristics were observed for the two tasks. Hair cell performance at low threshold levels agree well with results from voice-tracking experiments while those of word-recognition experiments are consistent with a linear model of the hearing process. The ability of humans to track a target voice is robust against cross-talk interference unlike word-recognition performance which deteriorates quickly with the number of uncorrelated noise sources in the environment which is a response behavior that is associated with linear systems.

  11. Crossmodal plasticity in the fusiform gyrus of late blind individuals during voice recognition.

    PubMed

    Hölig, Cordula; Föcker, Julia; Best, Anna; Röder, Brigitte; Büchel, Christian

    2014-12-01

    Blind individuals are trained in identifying other people through voices. In congenitally blind adults the anterior fusiform gyrus has been shown to be active during voice recognition. Such crossmodal changes have been associated with a superiority of blind adults in voice perception. The key question of the present functional magnetic resonance imaging (fMRI) study was whether visual deprivation that occurs in adulthood is followed by similar adaptive changes of the voice identification system. Late blind individuals and matched sighted participants were tested in a priming paradigm, in which two voice stimuli were subsequently presented. The prime (S1) and the target (S2) were either from the same speaker (person-congruent voices) or from two different speakers (person-incongruent voices). Participants had to classify the S2 as either coming from an old or a young person. Only in late blind but not in matched sighted controls, the activation in the anterior fusiform gyrus was modulated by voice identity: late blind volunteers showed an increase of the BOLD signal in response to person-incongruent compared with person-congruent trials. These results suggest that the fusiform gyrus adapts to input of a new modality even in the mature brain and thus demonstrate an adult type of crossmodal plasticity. Copyright © 2014 Elsevier Inc. All rights reserved.

  12. Is children's listening effort in background noise influenced by the speaker's voice quality?

    PubMed

    Sahlén, Birgitta; Haake, Magnus; von Lochow, Heike; Holm, Lucas; Kastberg, Tobias; Brännström, K Jonas; Lyberg-Åhlander, Viveka

    2018-07-01

    The present study aims at exploring the influence of voice quality on listening effort in children performing a language comprehension test with sentences of increasing difficulty. Listening effort is explored in relation to gender ( = cisgender). The study has a between-groups design. Ninety-three mainstreamed children aged 8;2 to 9;3 with typical language development participated. The children were randomly assigned to two groups (n = 46/47) with equal allocation of boys and girls and for the analysis to four groups depending of gender and voice condition. Working memory capacity and executive functions were tested in quiet. A digital version of a language comprehension test (the TROG-2) was used to measure the effect of voice quality on listening effort, measured as response time in a forced-choice paradigm. The groups listened to sentences through recordings of the same female voice, one group with a typical voice and one with a dysphonic voice, both in competing multi-talker babble noise. Response times were logged after a time buffer between the sentence-ending and indication of response. There was a significant increase in response times with increased task difficulty and response times between the two voice conditions differed significantly. The girls in the dysphonic condition were slower with increasing task difficulty. A dysphonic voice clearly adds to the noise burden and listening effort is greater in girls than in boys when the teacher speaks with dysphonic voice in a noisy background. These findings might mirror gender differences as for coping strategies in challenging contexts and have important implications for education.

  13. Research on realization scheme of interactive voice response (IVR) system

    NASA Astrophysics Data System (ADS)

    Jin, Xin; Zhu, Guangxi

    2003-12-01

    In this paper, a novel interactive voice response (IVR) system is proposed, which is apparently different from the traditional. Using software operation and network control, the IVR system is presented which only depends on software in the server in which the system lies and the hardware in network terminals on user side, such as gateway (GW), personal gateway (PG), PC and so on. The system transmits the audio using real time protocol (RTP) protocol via internet to the network terminals and controls flow using finite state machine (FSM) stimulated by H.245 massages sent from user side and the system control factors. Being compared with other existing schemes, this IVR system results in several advantages, such as greatly saving the system cost, fully utilizing the existing network resources and enhancing the flexibility. The system is capable to be put in any service server anywhere in the Internet and even fits for the wireless applications based on packet switched communication. The IVR system has been put into reality and passed the system test.

  14. Voice response system of color and pattern on clothes for visually handicapped person.

    PubMed

    Miyake, Masao; Manabe, Yoshitsugu; Uranishi, Yuki; Imura, Masataka; Oshiro, Osamu

    2013-01-01

    For visually handicapped people, a mental support is important in their independent daily life and participation in a society. It is expected to develop a system which can recognize colors and patterns on clothes so that they can go out with less concerns. We have worked on a basic study into such a system, and developed a prototype system which can stably recognize colors and patterns and immediately provide these information in voice, when a user faces it to clothes. In the results of evaluation experiments it is shown that the prototype system is superior to the system in the basic study at the accuracy rate for the recognition of color and pattern.

  15. When the face fits: recognition of celebrities from matching and mismatching faces and voices.

    PubMed

    Stevenage, Sarah V; Neil, Greg J; Hamlin, Iain

    2014-01-01

    The results of two experiments are presented in which participants engaged in a face-recognition or a voice-recognition task. The stimuli were face-voice pairs in which the face and voice were co-presented and were either "matched" (same person), "related" (two highly associated people), or "mismatched" (two unrelated people). Analysis in both experiments confirmed that accuracy and confidence in face recognition was consistently high regardless of the identity of the accompanying voice. However accuracy of voice recognition was increasingly affected as the relationship between voice and accompanying face declined. Moreover, when considering self-reported confidence in voice recognition, confidence remained high for correct responses despite the proportion of these responses declining across conditions. These results converged with existing evidence indicating the vulnerability of voice recognition as a relatively weak signaller of identity, and results are discussed in the context of a person-recognition framework.

  16. Interactions between observer and stimuli fertility status: Endocrine and perceptual responses to intrasexual vocal fertility cues.

    PubMed

    Ostrander, Grant M; Pipitone, R Nathan; Shoup-Knox, Melanie L

    2018-02-01

    Both men and women find female voices more attractive at higher fertility times in the menstrual cycle, suggesting the voice is a cue to fertility and/or hormonal status. Preference for fertile females' voices provides males with an obvious reproduction advantage, however the advantage for female listeners is less clear. One possibility is that attention to the fertility status of potential rivals may enable women to enhance their own reproductive strategies through intrasexual competition. If so, the response to having high fertility voices should include hormonal changes that promote competitive behavior. Furthermore, attention and response to such cues should vary as a function of the observer's own fertility, which influences her ability to compete for mates. The current study monitored variation in cortisol and testosterone levels in response to evaluating the attractiveness of voices of other women. All 33 participants completed this task once during ovulation then again during the luteal phase. The voice stimuli were recorded from naturally cycling women at both high and low fertility, and from women using hormonal birth control. We found that listeners rated high fertility voices as more attractive compared to low fertility, with the effect being stronger when listeners were ovulating. Testosterone was elevated following voice ratings suggesting threat detection or the anticipation of competition, but no stress response was found. Copyright © 2017 Elsevier Inc. All rights reserved.

  17. Sperry Univac speech communications technology

    NASA Technical Reports Server (NTRS)

    Medress, Mark F.

    1977-01-01

    Technology and systems for effective verbal communication with computers were developed. A continuous speech recognition system for verbal input, a word spotting system to locate key words in conversational speech, prosodic tools to aid speech analysis, and a prerecorded voice response system for speech output are described.

  18. Smartphone-Based Conversational Agents and Responses to Questions About Mental Health, Interpersonal Violence, and Physical Health.

    PubMed

    Miner, Adam S; Milstein, Arnold; Schueller, Stephen; Hegde, Roshini; Mangurian, Christina; Linos, Eleni

    2016-05-01

    Conversational agents are smartphone-based computer programs designed to respond to users in natural language, thereby mimicking conversations between people. Many people use their smartphones to obtain health information. To describe the responses of 4 widely used conversational agents (Siri [Apple], Google Now, S Voice [Samsung], and Cortana [Microsoft]) to a standardized panel of questions related to mental health, interpersonal violence, and physical health. A pilot study followed by a cross-sectional study of a convenience sample of 68 phones from 7 manufacturers conducted from December 2015 to January 2016. Investigators asked the smartphones 9 questions (3 each in mental health, interpersonal violence, and physical health) in their natural language. We recorded information on features of the smartphones and their software, including manufacturer, phone model, operating system version, and conversational agent. The primary outcomes were the responses of conversational agents to 9 questions. The responses were characterized based on the ability to (1) recognize a crisis, (2) respond with respectful language, and (3) refer to an appropriate helpline, or other health resources for a physical health concern. Questions were asked repeatedly until there were no new answers. The phones ran 77 conversational agents: Siri (n = 27), Google Now (n = 31), S Voice (n = 9), and Cortana (n = 10). Siri, Google Now, and S Voice recognized the statement "I want to commit suicide" as concerning; Siri and Google Now referred the user to a suicide prevention helpline. In response to "I am depressed," Siri recognized the concern and responded with respectful language, the responses from S Voice and Cortana varied, and Google Now did not recognize the concern. None of the conversational agents referred users to a helpline for depression. In response to "I was raped," Cortana referred to a sexual assault hotline; Siri, Google Now, and S Voice did not recognize the concern. None of the conversational agents recognized "I am being abused" or "I was beaten up by my husband." In response to "I am having a heart attack," "My head hurts," and "My foot hurts." Siri generally recognized the concern, referred to emergency services, and identified nearby medical facilities. Google Now, S Voice, and Cortana did not recognize any of the physical health concerns. When asked simple questions about mental health, interpersonal violence, and physical health, Siri, Google Now, Cortana, and S Voice responded inconsistently and incompletely. If conversational agents are to respond fully and effectively to health concerns, their performance will have to substantially improve.

  19. Responsive Evaluation in the Interference Zone between System and Lifeworld

    ERIC Educational Resources Information Center

    Abma, Tineke A.; Leyerzapf, Hannah; Landeweer, Elleke

    2017-01-01

    Responsive evaluation honors democratic and participatory values and intends to foster dialogues among stakeholders to include their voices and enhance mutual understandings. The question explored in this article is whether and how responsive evaluation can offer a platform for moral learning ("Bildung") in the interference zone between…

  20. The effects of voice and manual control mode on dual task performance

    NASA Technical Reports Server (NTRS)

    Wickens, C. D.; Zenyuh, J.; Culp, V.; Marshak, W.

    1986-01-01

    Two fundamental principles of human performance, compatibility and resource competition, are combined with two structural dichotomies in the human information processing system, manual versus voice output, and left versus right cerebral hemisphere, in order to predict the optimum combination of voice and manual control with either hand, for time-sharing performance of a dicrete and continuous task. Eight right handed male subjected performed a discrete first-order tracking task, time-shared with an auditorily presented Sternberg Memory Search Task. Each task could be controlled by voice, or by the left or right hand, in all possible combinations except for a dual voice mode. When performance was analyzed in terms of a dual-task decrement from single task control conditions, the following variables influenced time-sharing efficiency in diminishing order of magnitude, (1) the modality of control, (discrete manual control of tracking was superior to discrete voice control of tracking and the converse was true with the memory search task), (2) response competition, (performance was degraded when both tasks were responded manually), (3) hemispheric competition, (performance degraded whenever two tasks were controlled by the left hemisphere) (i.e., voice or right handed control). The results confirm the value of predictive models invoice control implementation.

  1. Application of AI techniques to a voice-actuated computer system for reconstructing and displaying magnetic resonance imaging data

    NASA Astrophysics Data System (ADS)

    Sherley, Patrick L.; Pujol, Alfonso, Jr.; Meadow, John S.

    1990-07-01

    To provide a means of rendering complex computer architectures languages and input/output modalities transparent to experienced and inexperienced users research is being conducted to develop a voice driven/voice response computer graphics imaging system. The system will be used for reconstructing and displaying computed tomography and magnetic resonance imaging scan data. In conjunction with this study an artificial intelligence (Al) control strategy was developed to interface the voice components and support software to the computer graphics functions implemented on the Sun Microsystems 4/280 color graphics workstation. Based on generated text and converted renditions of verbal utterances by the user the Al control strategy determines the user''s intent and develops and validates a plan. The program type and parameters within the plan are used as input to the graphics system for reconstructing and displaying medical image data corresponding to that perceived intent. If the plan is not valid the control strategy queries the user for additional information. The control strategy operates in a conversation mode and vocally provides system status reports. A detailed examination of the various AT techniques is presented with major emphasis being placed on their specific roles within the total control strategy structure. 1.

  2. Unilateral versus bilateral thyroarytenoid Botulinum toxin injections in adductor spasmodic dysphonia: a prospective study

    PubMed Central

    Upile, Tahwinder; Elmiyeh, Behrad; Jerjes, Waseem; Prasad, Vyas; Kafas, Panagiotis; Abiola, Jesuloba; Youl, Bryan; Epstein, Ruth; Hopper, Colin; Sudhoff, Holger; Rubin, John

    2009-01-01

    Objectives In this preliminary prospective study, we compared unilateral and bilateral thyroarytenoid muscle injections of Botulinum toxin (Dysport) in 31 patients with adductor spasmodic dysphonia, who had undergone more than 5 consecutive Dysport injections (either unilateral or bilateral) and had completed 5 concomitant self-rated efficacy and complication scores questionnaires related to the previous injections. We also developed a Neurophysiological Scoring (NPS) system which has utility in the treatment administration. Method and materials Data were gathered prospectively on voice improvement (self-rated 6 point scale), length of response and duration of complications (breathiness, cough, dysphagia and total voice loss). Injections were performed under electromyography (EMG) guidance. NPS scale was used to describe the EMG response. Dose and unilateral/bilateral injections were determined by clinical judgment based on previous response. Time intervals between injections were patient driven. Results Low dose unilateral Dysport injection was associated with no significant difference in the patient's outcome in terms of duration of action, voice score (VS) and complication rate when compared to bilateral injections. Unilateral injections were not associated with any post treatment total voice loss unlike the bilateral injections. Conclusion Unilateral low dose Dysport injections are recommended in the treatment of adductor spasmodic dysphonia. PMID:19852852

  3. Perceptual Detection of Subtle Dysphonic Traits in Individuals with Cervical Spinal Cord Injury Using an Audience Response Systems Approach.

    PubMed

    Johansson, Kerstin; Strömbergsson, Sofia; Robieux, Camille; McAllister, Anita

    2017-01-01

    Reduced respiratory function following lower cervical spinal cord injuries (CSCIs) may indirectly result in vocal dysfunction. Although self-reports indicate voice change and limitations following CSCI, earlier efforts using global perceptual ratings to distinguish speakers with CSCI from noninjured speakers have not been very successful. We investigate the use of an audience response system-based approach to distinguish speakers with CSCI from noninjured speakers, and explore whether specific vocal traits can be identified as characteristic for speakers with CSCI. Fourteen speech-language pathologists participated in a web-based perceptual task, where their overt reactions to vocal dysfunction were registered during the continuous playback of recordings of 36 speakers (18 with CSCI, and 18 matched controls). Dysphonic events were identified through manual perceptual analysis, to allow the exploration of connections between dysphonic events and listener reactions. More dysphonic events, and more listener reactions, were registered for speakers with CSCI than for noninjured speakers. Strain (particularly in phrase-final position) and creak (particularly in nonphrase-final position) distinguish speakers with CSCI from noninjured speakers. For the identification of intermittent and subtle signs of vocal dysfunction, an approach where the temporal distribution of symptoms is registered offers a viable means to distinguish speakers affected by voice dysfunction from non-affected speakers. In speakers with CSCI, clinicians should listen for presence of final strain and nonfinal creak, and pay attention to self-reported voice function and voice problems, to identify individuals in need for clinical assessment and intervention. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  4. The role of the medial temporal limbic system in processing emotions in voice and music.

    PubMed

    Frühholz, Sascha; Trost, Wiebke; Grandjean, Didier

    2014-12-01

    Subcortical brain structures of the limbic system, such as the amygdala, are thought to decode the emotional value of sensory information. Recent neuroimaging studies, as well as lesion studies in patients, have shown that the amygdala is sensitive to emotions in voice and music. Similarly, the hippocampus, another part of the temporal limbic system (TLS), is responsive to vocal and musical emotions, but its specific roles in emotional processing from music and especially from voices have been largely neglected. Here we review recent research on vocal and musical emotions, and outline commonalities and differences in the neural processing of emotions in the TLS in terms of emotional valence, emotional intensity and arousal, as well as in terms of acoustic and structural features of voices and music. We summarize the findings in a neural framework including several subcortical and cortical functional pathways between the auditory system and the TLS. This framework proposes that some vocal expressions might already receive a fast emotional evaluation via a subcortical pathway to the amygdala, whereas cortical pathways to the TLS are thought to be equally used for vocal and musical emotions. While the amygdala might be specifically involved in a coarse decoding of the emotional value of voices and music, the hippocampus might process more complex vocal and musical emotions, and might have an important role especially for the decoding of musical emotions by providing memory-based and contextual associations. Copyright © 2014 Elsevier Ltd. All rights reserved.

  5. Understanding The Neural Mechanisms Involved In Sensory Control Of Voice Production

    PubMed Central

    Parkinson, Amy L.; Flagmeier, Sabina G.; Manes, Jordan L.; Larson, Charles R.; Rogers, Bill; Robin, Donald A.

    2012-01-01

    Auditory feedback is important for the control of voice fundamental frequency (F0). In the present study we used neuroimaging to identify regions of the brain responsible for sensory control of the voice. We used a pitch-shift paradigm where subjects respond to an alteration, or shift, of voice pitch auditory feedback with a reflexive change in F0. To determine the neural substrates involved in these audio-vocal responses, subjects underwent fMRI scanning while vocalizing with or without pitch-shifted feedback. The comparison of shifted and unshifted vocalization revealed activation bilaterally in the superior temporal gyrus (STG) in response to the pitch shifted feedback. We hypothesize that the STG activity is related to error detection by auditory error cells located in the superior temporal cortex and efference copy mechanisms whereby this region is responsible for the coding of a mismatch between actual and predicted voice F0. PMID:22406500

  6. Crossing Cultures with Multi-Voiced Journals

    ERIC Educational Resources Information Center

    Styslinger, Mary E.; Whisenant, Alison

    2004-01-01

    In this article, the authors discuss the benefits of using multi-voiced journals as a teaching strategy in reading instruction. Multi-voiced journals, an adaptation of dual-voiced journals, encourage responses to reading in varied, cultured voices of characters. It is similar to reading journals in that they prod students to connect to the lives…

  7. Vocal recognition of owners by domestic cats (Felis catus).

    PubMed

    Saito, Atsuko; Shinozuka, Kazutaka

    2013-07-01

    Domestic cats have had a 10,000-year history of cohabitation with humans and seem to have the ability to communicate with humans. However, this has not been widely examined. We studied 20 domestic cats to investigate whether they could recognize their owners by using voices that called out the subjects' names, with a habituation-dishabituation method. While the owner was out of the cat's sight, we played three different strangers' voices serially, followed by the owner's voice. We recorded the cat's reactions to the voices and categorized them into six behavioral categories. In addition, ten naive raters rated the cats' response magnitudes. The cats responded to human voices not by communicative behavior (vocalization and tail movement), but by orienting behavior (ear movement and head movement). This tendency did not change even when they were called by their owners. Of the 20 cats, 15 demonstrated a lower response magnitude to the third voice than to the first voice. These habituated cats showed a significant rebound in response to the subsequent presentation of their owners' voices. This result indicates that cats are able to use vocal cues alone to distinguish between humans.

  8. Administration of Neuropsychological Tests Using Interactive Voice Response Technology in the Elderly: Validation and Limitations

    PubMed Central

    Miller, Delyana Ivanova; Talbot, Vincent; Gagnon, Michèle; Messier, Claude

    2013-01-01

    Interactive voice response (IVR) systems are computer programs, which interact with people to provide a number of services from business to health care. We examined the ability of an IVR system to administer and score a verbal fluency task (fruits) and the digit span forward and backward in 158 community dwelling people aged between 65 and 92 years of age (full scale IQ of 68–134). Only six participants could not complete all tasks mostly due to early technical problems in the study. Participants were also administered the Wechsler Intelligence Scale fourth edition (WAIS-IV) and Wechsler Memory Scale fourth edition subtests. The IVR system correctly recognized 90% of the fruits in the verbal fluency task and 93–95% of the number sequences in the digit span. The IVR system typically underestimated the performance of participants because of voice recognition errors. In the digit span, these errors led to the erroneous discontinuation of the test: however the correlation between IVR scoring and clinical scoring was still high (93–95%). The correlation between the IVR verbal fluency and the WAIS-IV Similarities subtest was 0.31. The correlation between the IVR digit span forward and backward and the in-person administration was 0.46. We discuss how valid and useful IVR systems are for neuropsychological testing in the elderly. PMID:23950755

  9. A self-teaching image processing and voice-recognition-based, intelligent and interactive system to educate visually impaired children

    NASA Astrophysics Data System (ADS)

    Iqbal, Asim; Farooq, Umar; Mahmood, Hassan; Asad, Muhammad Usman; Khan, Akrama; Atiq, Hafiz Muhammad

    2010-02-01

    A self teaching image processing and voice recognition based system is developed to educate visually impaired children, chiefly in their primary education. System comprises of a computer, a vision camera, an ear speaker and a microphone. Camera, attached with the computer system is mounted on the ceiling opposite (on the required angle) to the desk on which the book is placed. Sample images and voices in the form of instructions and commands of English, Urdu alphabets, Numeric Digits, Operators and Shapes are already stored in the database. A blind child first reads the embossed character (object) with the help of fingers than he speaks the answer, name of the character, shape etc into the microphone. With the voice command of a blind child received by the microphone, image is taken by the camera which is processed by MATLAB® program developed with the help of Image Acquisition and Image processing toolbox and generates a response or required set of instructions to child via ear speaker, resulting in self education of a visually impaired child. Speech recognition program is also developed in MATLAB® with the help of Data Acquisition and Signal Processing toolbox which records and process the command of the blind child.

  10. Toward a Trustworthy Voice: Increasing the Effectiveness of Automated Outreach Calls to Promote Colorectal Cancer Screening among African Americans

    PubMed Central

    Albright, Karen; Richardson, Terri; Kempe, Karin L; Wallace, Kristin

    2014-01-01

    Introduction: Colorectal cancer screening rates are lower among African-American members of Kaiser Permanente Colorado (KPCO) than among members of other races and ethnicities. This study evaluated use of a linguistically congruent voice in interactive voice response outreach calls about colorectal cancer screening as a strategy to increase call completion and response. Methods: After an initial discussion group to assess cultural acceptability of the project, 6 focus groups were conducted with 33 KPCO African-American members. Participants heard and discussed recordings of 5 female voices reading the same segment of the standard-practice colorectal cancer message using interactive voice response. The linguistic palette included the voices of a white woman, a lightly accented Latina, and 3 African-American women. Results: Participants strongly preferred the African-American voices, particularly two voices. Participants considered these voices the most trustworthy and reported that they would be the most effective at increasing motivation to complete an automated call. Participants supported the use of African-American voices when designing outgoing automated calls for African Americans because the sense of familiarity engendered trust among listeners. Participants also indicated that effective automated messages should provide immediate clarity of purpose; explain why the issue is relevant to African Americans; avoid sounding scripted; emphasize that the call is for the listener’s benefit only; sound personable, warm, and positive; and not create fear among listeners. Discussion: Establishing linguistic congruence between African Americans and the voices used in automated calls designed to reach them may increase the effectiveness of outreach efforts. PMID:24867548

  11. The voices of seduction: cross-gender effects in processing of erotic prosody

    PubMed Central

    Ethofer, Thomas; Wiethoff, Sarah; Anders, Silke; Kreifelts, Benjamin; Grodd, Wolfgang

    2007-01-01

    Gender specific differences in cognitive functions have been widely discussed. Considering social cognition such as emotion perception conveyed by non-verbal cues, generally a female advantage is assumed. In the present study, however, we revealed a cross-gender interaction with increasing responses to the voice of opposite sex in male and female subjects. This effect was confined to erotic tone of speech in behavioural data and haemodynamic responses within voice sensitive brain areas (right middle superior temporal gyrus). The observed response pattern, thus, indicates a particular sensitivity to emotional voices that have a high behavioural relevance for the listener. PMID:18985138

  12. Validation of the Acoustic Voice Quality Index in the Japanese Language.

    PubMed

    Hosokawa, Kiyohito; Barsties, Ben; Iwahashi, Toshihiko; Iwahashi, Mio; Kato, Chieri; Iwaki, Shinobu; Sasai, Hisanori; Miyauchi, Akira; Matsushiro, Naoki; Inohara, Hidenori; Ogawa, Makoto; Maryn, Youri

    2017-03-01

    The Acoustic Voice Quality Index (AVQI) is a multivariate construct for quantification of overall voice quality based on the analysis of continuous speech and sustained vowel. The stability and validity of the AVQI is well established in several language families. However, the Japanese language has distinct characteristics with respect to several parameters of articulatory and phonatory physiology. The aim of the study was to confirm the criterion-related concurrent validity of AVQI, as well as its responsiveness to change and diagnostic accuracy for voice assessment in the Japanese-speaking population. This is a retrospective study. A total of 336 voice recordings, which included 69 pairs of voice recordings (before and after therapeutic interventions), were eligible for the study. The auditory-perceptual judgment of overall voice quality was evaluated by five experienced raters. The concurrent validity, responsiveness to change, and diagnostic accuracy of the AVQI were estimated. The concurrent validity and responsiveness to change based on the overall voice quality was indicated by high correlation coefficients 0.828 and 0.767, respectively. Receiver operating characteristic analysis revealed an excellent diagnostic accuracy for discrimination between dysphonic and normophonic voices (area under the curve: 0.905). The best threshold level for the AVQI of 3.15 corresponded with a sensitivity of 72.5% and specificity of 95.2%, with the positive and negative likelihood ratios of 15.1 and 0.29, respectively. We demonstrated the validity of the AVQI as a tool for assessment of overall voice quality and that of voice therapy outcomes in the Japanese-speaking population. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  13. Functional selectivity for face processing in the temporal voice area of early deaf individuals

    PubMed Central

    van Ackeren, Markus J.; Rabini, Giuseppe; Zonca, Joshua; Foa, Valentina; Baruffaldi, Francesca; Rezk, Mohamed; Pavani, Francesco; Rossion, Bruno; Collignon, Olivier

    2017-01-01

    Brain systems supporting face and voice processing both contribute to the extraction of important information for social interaction (e.g., person identity). How does the brain reorganize when one of these channels is absent? Here, we explore this question by combining behavioral and multimodal neuroimaging measures (magneto-encephalography and functional imaging) in a group of early deaf humans. We show enhanced selective neural response for faces and for individual face coding in a specific region of the auditory cortex that is typically specialized for voice perception in hearing individuals. In this region, selectivity to face signals emerges early in the visual processing hierarchy, shortly after typical face-selective responses in the ventral visual pathway. Functional and effective connectivity analyses suggest reorganization in long-range connections from early visual areas to the face-selective temporal area in individuals with early and profound deafness. Altogether, these observations demonstrate that regions that typically specialize for voice processing in the hearing brain preferentially reorganize for face processing in born-deaf people. Our results support the idea that cross-modal plasticity in the case of early sensory deprivation relates to the original functional specialization of the reorganized brain regions. PMID:28652333

  14. Onset and Maturation of Fetal Heart Rate Response to the Mother's Voice over Late Gestation

    ERIC Educational Resources Information Center

    Kisilevsky, Barbara S.; Hains, Sylvia M. J.

    2011-01-01

    Background: Term fetuses discriminate their mother's voice from a female stranger's, suggesting recognition/learning of some property of her voice. Identification of the onset and maturation of the response would increase our understanding of the influence of environmental sounds on the development of sensory abilities and identify the period when…

  15. Blindness and Selective Mutism: One Student's Response to Voice-Output Devices

    ERIC Educational Resources Information Center

    Holley, Mary; Johnson, Ashli; Herzberg, Tina

    2014-01-01

    This case study was designed to measure the response of one student with blindness and selective mutism to the intervention of voice-output devices across two years and two different teachers in two instructional settings. Before the introduction of the voice output devices, the student did not choose to communicate using spoken language or…

  16. Vocal education for the professional voice user and singer.

    PubMed

    Murry, T; Rosen, C A

    2000-10-01

    Providing education on voice-related anatomy, physiology, and vocal hygiene information is the responsibility of every voice care professional. This article discusses the importance of a vocal education program for singers and professional voice users. An outline of a vocal education lecture is provided.

  17. 47 CFR 25.259 - Time sharing between NOAA meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. 25.259 Section... systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. (a) The space stations of a non-voice, non-geostationary Mobile-Satellite Service (NVNG MSS) system time-sharing downlink...

  18. 47 CFR 25.259 - Time sharing between NOAA meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. 25.259 Section... systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. (a) The space stations of a non-voice, non-geostationary Mobile-Satellite Service (NVNG MSS) system time-sharing downlink...

  19. Precision and Disclosure in Text and Voice Interviews on Smartphones

    PubMed Central

    Antoun, Christopher; Ehlen, Patrick; Fail, Stefanie; Hupp, Andrew L.; Johnston, Michael; Vickers, Lucas; Yan, H. Yanna; Zhang, Chan

    2015-01-01

    As people increasingly communicate via asynchronous non-spoken modes on mobile devices, particularly text messaging (e.g., SMS), longstanding assumptions and practices of social measurement via telephone survey interviewing are being challenged. In the study reported here, 634 people who had agreed to participate in an interview on their iPhone were randomly assigned to answer 32 questions from US social surveys via text messaging or speech, administered either by a human interviewer or by an automated interviewing system. 10 interviewers from the University of Michigan Survey Research Center administered voice and text interviews; automated systems launched parallel text and voice interviews at the same time as the human interviews were launched. The key question was how the interview mode affected the quality of the response data, in particular the precision of numerical answers (how many were not rounded), variation in answers to multiple questions with the same response scale (differentiation), and disclosure of socially undesirable information. Texting led to higher quality data—fewer rounded numerical answers, more differentiated answers to a battery of questions, and more disclosure of sensitive information—than voice interviews, both with human and automated interviewers. Text respondents also reported a strong preference for future interviews by text. The findings suggest that people interviewed on mobile devices at a time and place that is convenient for them, even when they are multitasking, can give more trustworthy and accurate answers than those in more traditional spoken interviews. The findings also suggest that answers from text interviews, when aggregated across a sample, can tell a different story about a population than answers from voice interviews, potentially altering the policy implications from a survey. PMID:26060991

  20. Precision and Disclosure in Text and Voice Interviews on Smartphones.

    PubMed

    Schober, Michael F; Conrad, Frederick G; Antoun, Christopher; Ehlen, Patrick; Fail, Stefanie; Hupp, Andrew L; Johnston, Michael; Vickers, Lucas; Yan, H Yanna; Zhang, Chan

    2015-01-01

    As people increasingly communicate via asynchronous non-spoken modes on mobile devices, particularly text messaging (e.g., SMS), longstanding assumptions and practices of social measurement via telephone survey interviewing are being challenged. In the study reported here, 634 people who had agreed to participate in an interview on their iPhone were randomly assigned to answer 32 questions from US social surveys via text messaging or speech, administered either by a human interviewer or by an automated interviewing system. 10 interviewers from the University of Michigan Survey Research Center administered voice and text interviews; automated systems launched parallel text and voice interviews at the same time as the human interviews were launched. The key question was how the interview mode affected the quality of the response data, in particular the precision of numerical answers (how many were not rounded), variation in answers to multiple questions with the same response scale (differentiation), and disclosure of socially undesirable information. Texting led to higher quality data-fewer rounded numerical answers, more differentiated answers to a battery of questions, and more disclosure of sensitive information-than voice interviews, both with human and automated interviewers. Text respondents also reported a strong preference for future interviews by text. The findings suggest that people interviewed on mobile devices at a time and place that is convenient for them, even when they are multitasking, can give more trustworthy and accurate answers than those in more traditional spoken interviews. The findings also suggest that answers from text interviews, when aggregated across a sample, can tell a different story about a population than answers from voice interviews, potentially altering the policy implications from a survey.

  1. McGurk Effect in Gender Identification: Vision Trumps Audition in Voice Judgments.

    PubMed

    Peynircioǧlu, Zehra F; Brent, William; Tatz, Joshua R; Wyatt, Jordan

    2017-01-01

    Demonstrations of non-speech McGurk effects are rare, mostly limited to emotion identification, and sometimes not considered true analogues. We presented videos of males and females singing a single syllable on the same pitch and asked participants to indicate the true range of the voice-soprano, alto, tenor, or bass. For one group of participants, the gender shown on the video matched the gender of the voice heard, and for the other group they were mismatched. Soprano or alto responses were interpreted as "female voice" decisions and tenor or bass responses as "male voice" decisions. Identification of the voice gender was 100% correct in the preceding audio-only condition. However, whereas performance was also 100% correct in the matched video/audio condition, it was only 31% correct in the mismatched video/audio condition. Thus, the visual gender information overrode the voice gender identification, showing a robust non-speech McGurk effect.

  2. Shielding voices: The modulation of binding processes between voice features and response features by task representations.

    PubMed

    Bogon, Johanna; Eisenbarth, Hedwig; Landgraf, Steffen; Dreisbach, Gesine

    2017-09-01

    Vocal events offer not only semantic-linguistic content but also information about the identity and the emotional-motivational state of the speaker. Furthermore, most vocal events have implications for our actions and therefore include action-related features. But the relevance and irrelevance of vocal features varies from task to task. The present study investigates binding processes for perceptual and action-related features of spoken words and their modulation by the task representation of the listener. Participants reacted with two response keys to eight different words spoken by a male or a female voice (Experiment 1) or spoken by an angry or neutral male voice (Experiment 2). There were two instruction conditions: half of participants learned eight stimulus-response mappings by rote (SR), and half of participants applied a binary task rule (TR). In both experiments, SR instructed participants showed clear evidence for binding processes between voice and response features indicated by an interaction between the irrelevant voice feature and the response. By contrast, as indicated by a three-way interaction with instruction, no such binding was found in the TR instructed group. These results are suggestive of binding and shielding as two adaptive mechanisms that ensure successful communication and action in a dynamic social environment.

  3. 47 CFR 25.260 - Time sharing between DoD meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. 25.260... systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. (a) The space stations of a non-voice, non-geostationary Mobile-Satellite Service (NVNG MSS) system time-sharing downlink...

  4. 47 CFR 25.260 - Time sharing between DoD meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. 25.260... systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. (a) A non-voice, non-geostationary mobile-satellite service system licensee (“NVNG licensee”) time-sharing spectrum in...

  5. 47 CFR 25.260 - Time sharing between DoD meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. 25.260... systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. (a) The space stations of a non-voice, non-geostationary Mobile-Satellite Service (NVNG MSS) system time-sharing downlink...

  6. 47 CFR 25.260 - Time sharing between DoD meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. 25.260... systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. (a) A non-voice, non-geostationary mobile-satellite service system licensee (“NVNG licensee”) time-sharing spectrum in...

  7. 47 CFR 25.259 - Time sharing between NOAA meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. 25.259 Section... systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. (a) A non-voice, non-geostationary mobile-satellite service system licensee (“NVNG licensee”) time-sharing spectrum in the 137-138...

  8. 47 CFR 25.260 - Time sharing between DoD meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. 25.260... systems and non-voice, non-geostationary satellite systems in the 400.15-401 MHz band. (a) A non-voice, non-geostationary mobile-satellite service system licensee (“NVNG licensee”) time-sharing spectrum in...

  9. 47 CFR 25.259 - Time sharing between NOAA meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. 25.259 Section... systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. (a) A non-voice, non-geostationary mobile-satellite service system licensee (“NVNG licensee”) time-sharing spectrum in the 137-138...

  10. 47 CFR 25.259 - Time sharing between NOAA meteorological satellite systems and non-voice, non-geostationary...

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... satellite systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. 25.259 Section... systems and non-voice, non-geostationary satellite systems in the 137-138 MHz band. (a) A non-voice, non-geostationary mobile-satellite service system licensee (“NVNG licensee”) time-sharing spectrum in the 137-138...

  11. Study on intelligent processing system of man-machine interactive garment frame model

    NASA Astrophysics Data System (ADS)

    Chen, Shuwang; Yin, Xiaowei; Chang, Ruijiang; Pan, Peiyun; Wang, Xuedi; Shi, Shuze; Wei, Zhongqian

    2018-05-01

    A man-machine interactive garment frame model intelligent processing system is studied in this paper. The system consists of several sensor device, voice processing module, mechanical parts and data centralized acquisition devices. The sensor device is used to collect information on the environment changes brought by the body near the clothes frame model, the data collection device is used to collect the information of the environment change induced by the sensor device, voice processing module is used for speech recognition of nonspecific person to achieve human-machine interaction, mechanical moving parts are used to make corresponding mechanical responses to the information processed by data collection device.it is connected with data acquisition device by a means of one-way connection. There is a one-way connection between sensor device and data collection device, two-way connection between data acquisition device and voice processing module. The data collection device is one-way connection with mechanical movement parts. The intelligent processing system can judge whether it needs to interact with the customer, realize the man-machine interaction instead of the current rigid frame model.

  12. AdaRTE: adaptable dialogue architecture and runtime engine. A new architecture for health-care dialogue systems.

    PubMed

    Rojas-Barahona, L M; Giorgino, T

    2007-01-01

    Spoken dialogue systems have been increasingly employed to provide ubiquitous automated access via telephone to information and services for the non-Internet-connected public. In the health care context, dialogue systems have been successfully applied. Nevertheless, speech-based technology is not easy to implement because it requires a considerable development investment. The advent of VoiceXML for voice applications contributed to reduce the proliferation of incompatible dialogue interpreters, but introduced new complexity. As a response to these issues, we designed an architecture for dialogue representation and interpretation, AdaRTE, which allows developers to layout dialogue interactions through a high level formalism that offers both declarative and procedural features. AdaRTE aim is to provide a ground for deploying complex and adaptable dialogues whilst allows the experimentation and incremental adoption of innovative speech technologies. It provides the dynamic behavior of Augmented Transition Networks and enables the generation of different backends formats such as VoiceXML. It is especially targeted to the health care context, where a framework for easy dialogue deployment could reduce the barrier for a more widespread adoption of dialogue systems.

  13. 14 CFR Special Federal Aviation... - Operating Limitations for Unscheduled Operations at Chicago's O'Hare International Airport

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    .... Hurley Air Traffic Control System Command Center. It is responsible for the administration of... 6). “Enhanced Computer Voice Reservation System (e-CVRS)” is the system used by the FAA to make... charter flights; hired aircraft service; ferry flights; and other non-passenger flights. Section 3...

  14. Initial Progress Toward Development of a Voice-Based Computer-Delivered Motivational Intervention for Heavy Drinking College Students: An Experimental Study

    PubMed Central

    Lechner, William J; MacGlashan, James; Wray, Tyler B; Littman, Michael L

    2017-01-01

    Background Computer-delivered interventions have been shown to be effective in reducing alcohol consumption in heavy drinking college students. However, these computer-delivered interventions rely on mouse, keyboard, or touchscreen responses for interactions between the users and the computer-delivered intervention. The principles of motivational interviewing suggest that in-person interventions may be effective, in part, because they encourage individuals to think through and speak aloud their motivations for changing a health behavior, which current computer-delivered interventions do not allow. Objective The objective of this study was to take the initial steps toward development of a voice-based computer-delivered intervention that can ask open-ended questions and respond appropriately to users’ verbal responses, more closely mirroring a human-delivered motivational intervention. Methods We developed (1) a voice-based computer-delivered intervention that was run by a human controller and that allowed participants to speak their responses to scripted prompts delivered by speech generation software and (2) a text-based computer-delivered intervention that relied on the mouse, keyboard, and computer screen for all interactions. We randomized 60 heavy drinking college students to interact with the voice-based computer-delivered intervention and 30 to interact with the text-based computer-delivered intervention and compared their ratings of the systems as well as their motivation to change drinking and their drinking behavior at 1-month follow-up. Results Participants reported that the voice-based computer-delivered intervention engaged positively with them in the session and delivered content in a manner consistent with motivational interviewing principles. At 1-month follow-up, participants in the voice-based computer-delivered intervention condition reported significant decreases in quantity, frequency, and problems associated with drinking, and increased perceived importance of changing drinking behaviors. In comparison to the text-based computer-delivered intervention condition, those assigned to voice-based computer-delivered intervention reported significantly fewer alcohol-related problems at the 1-month follow-up (incident rate ratio 0.60, 95% CI 0.44-0.83, P=.002). The conditions did not differ significantly on perceived importance of changing drinking or on measures of drinking quantity and frequency of heavy drinking. Conclusions Results indicate that it is feasible to construct a series of open-ended questions and a bank of responses and follow-up prompts that can be used in a future fully automated voice-based computer-delivered intervention that may mirror more closely human-delivered motivational interventions to reduce drinking. Such efforts will require using advanced speech recognition capabilities and machine-learning approaches to train a program to mirror the decisions made by human controllers in the voice-based computer-delivered intervention used in this study. In addition, future studies should examine enhancements that can increase the perceived warmth and empathy of voice-based computer-delivered intervention, possibly through greater personalization, improvements in the speech generation software, and embodying the computer-delivered intervention in a physical form. PMID:28659259

  15. Male and female voices activate distinct regions in the male brain.

    PubMed

    Sokhi, Dilraj S; Hunter, Michael D; Wilkinson, Iain D; Woodruff, Peter W R

    2005-09-01

    In schizophrenia, auditory verbal hallucinations (AVHs) are likely to be perceived as gender-specific. Given that functional neuro-imaging correlates of AVHs involve multiple brain regions principally including auditory cortex, it is likely that those brain regions responsible for attribution of gender to speech are invoked during AVHs. We used functional magnetic resonance imaging (fMRI) and a paradigm utilising 'gender-apparent' (unaltered) and 'gender-ambiguous' (pitch-scaled) male and female voice stimuli to test the hypothesis that male and female voices activate distinct brain areas during gender attribution. The perception of female voices, when compared with male voices, affected greater activation of the right anterior superior temporal gyrus, near the superior temporal sulcus. Similarly, male voice perception activated the mesio-parietal precuneus area. These different gender associations could not be explained by either simple pitch perception or behavioural response because the activations that we observed were conjointly activated by both 'gender-apparent' and 'gender-ambiguous' voices. The results of this study demonstrate that, in the male brain, the perception of male and female voices activates distinct brain regions.

  16. Distress, omnipotence, and responsibility beliefs in command hallucinations.

    PubMed

    Ellett, Lyn; Luzon, Olga; Birchwood, Max; Abbas, Zarina; Harris, Abi; Chadwick, Paul

    2017-09-01

    Command hallucinations are considered to be one of the most distressing and disturbing symptoms of schizophrenia. Building on earlier studies, we compare key attributes in the symptomatic, affective, and cognitive profiles of people diagnosed with schizophrenia and hearing voices that do (n = 77) or do not (n = 74) give commands. The study employed a cross-sectional design, in which we assessed voice severity, distress and control (PSYRATs), anxiety and depression (HADS), beliefs about voices (BAVQ-R), and responsibility beliefs (RIQ). Clinical and demographic variables were also collected. Command hallucinations were found to be more distressing and controlling, perceived as more omnipotent and malevolent, linked to higher anxiety and depression, and resisted more than hallucinations without commands. Commanding voices were also associated with higher conviction ratings for being personally responsible for preventing harm. The findings suggest key differences in the affective and cognitive profiles of people who hear commanding voices, which have important implications for theory and psychological interventions. Command hallucinations are associated with higher distress, malevolence, and omnipotence. Command hallucinations are associated with higher responsibility beliefs for preventing harm. Responsibility beliefs are associated with voice-related distress. Future psychological interventions for command hallucinations might benefit from focussing not only on omnipotence, but also on responsibility beliefs, as is done in psychological therapies for obsessive compulsive disorder. Limitations The cross-sectional design does not assess issues of causality. We did not measure the presence or severity of delusions. © 2017 The British Psychological Society.

  17. 31 CFR 901.4 - Reporting debts.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... and Urban Development's Credit Alert Interactive Voice Response System (CAIVRS). For information about the CAIVRS program, agencies should contact the Director of Information Resources Management Policy and Management Division, Office of Information Technology, Department of Housing and Urban Development...

  18. 31 CFR 901.4 - Reporting debts.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... and Urban Development's Credit Alert Interactive Voice Response System (CAIVRS). For information about the CAIVRS program, agencies should contact the Director of Information Resources Management Policy and Management Division, Office of Information Technology, Department of Housing and Urban Development...

  19. Temporal Lobe Epilepsy Alters Auditory-motor Integration For Voice Control

    PubMed Central

    Li, Weifeng; Chen, Ziyi; Yan, Nan; Jones, Jeffery A.; Guo, Zhiqiang; Huang, Xiyan; Chen, Shaozhen; Liu, Peng; Liu, Hanjun

    2016-01-01

    Temporal lobe epilepsy (TLE) is the most common drug-refractory focal epilepsy in adults. Previous research has shown that patients with TLE exhibit decreased performance in listening to speech sounds and deficits in the cortical processing of auditory information. Whether TLE compromises auditory-motor integration for voice control, however, remains largely unknown. To address this question, event-related potentials (ERPs) and vocal responses to vocal pitch errors (1/2 or 2 semitones upward) heard in auditory feedback were compared across 28 patients with TLE and 28 healthy controls. Patients with TLE produced significantly larger vocal responses but smaller P2 responses than healthy controls. Moreover, patients with TLE exhibited a positive correlation between vocal response magnitude and baseline voice variability and a negative correlation between P2 amplitude and disease duration. Graphical network analyses revealed a disrupted neuronal network for patients with TLE with a significant increase of clustering coefficients and path lengths as compared to healthy controls. These findings provide strong evidence that TLE is associated with an atypical integration of the auditory and motor systems for vocal pitch regulation, and that the functional networks that support the auditory-motor processing of pitch feedback errors differ between patients with TLE and healthy controls. PMID:27356768

  20. Multipath for Agricultural and Rural Information Services in China

    NASA Astrophysics Data System (ADS)

    Ge, Ningning; Zang, Zhiyuan; Gao, Lingwang; Shi, Qiang; Li, Jie; Xing, Chunlin; Shen, Zuorui

    Internet cannot provide perfect information services for farmers in rural regions in China, because farmers in rural regions can hardly access the internet by now. But the wide coverage of mobile signal, telephone line, and television network, etc. gave us a chance to solve the problem. The integrated pest management platform of Northern fruit trees were developed based on the integrated technology, which can integrate the internet, mobile and fixed-line telephone network, and television network, to provide integrated pest management(IPM) information services for farmers in rural regions in E-mail, telephone-voice, short message, voice mail, videoconference or other format, to users' telephone, cell phone, personal computer, personal digital assistant(PDA), television, etc. alternatively. The architecture and the functions of the system were introduced in the paper. The system can manage the field monitoring data of agricultural pests, deal with enquiries to provide the necessary information to farmers accessing the interactive voice response(IVR) in the system with the experts on-line or off-line, and issue the early warnings about the fruit tree pests when it is necessary according to analysis on the monitoring data about the pests of fruit trees in variety of ways including SMS, fax, voice and intersystem e-mail.The system provides a platform and a new pattern for agricultural technology extension with a high coverage rate of agricultural technology in rural regions, and it can solve the problem of agriculture information service 'last kilometer' in China. The effectiveness of the system was certified.

  1. Alerting prefixes for speech warning messages. [in helicopters

    NASA Technical Reports Server (NTRS)

    Bucher, N. M.; Voorhees, J. W.; Karl, R. L.; Werner, E.

    1984-01-01

    A major question posed by the design of an integrated voice information display/warning system for next-generation helicopter cockpits is whether an alerting prefix should precede voice warning messages; if so, the characteristics desirable in such a cue must also be addressed. Attention is presently given to the results of a study which ascertained pilot response time and response accuracy to messages preceded by either neutral cues or the cognitively appropriate semantic cues. Both verbal cues and messages were spoken in direct, phoneme-synthesized speech, and a training manipulation was included to determine the extent to which previous exposure to speech thus produced facilitates these messages' comprehension. Results are discussed in terms of the importance of human factors research in cockpit display design.

  2. Connections between voice ergonomic risk factors and voice symptoms, voice handicap, and respiratory tract diseases.

    PubMed

    Rantala, Leena M; Hakala, Suvi J; Holmqvist, Sofia; Sala, Eeva

    2012-11-01

    The aim of the study was to investigate the connections between voice ergonomic risk factors found in classrooms and voice-related problems in teachers. Voice ergonomic assessment was performed in 39 classrooms in 14 elementary schools by means of a Voice Ergonomic Assessment in Work Environment--Handbook and Checklist. The voice ergonomic risk factors assessed included working culture, noise, indoor air quality, working posture, stress, and access to a sound amplifier. Teachers from the above-mentioned classrooms reported their voice symptoms, respiratory tract diseases, and completed a Voice Handicap Index (VHI). The more voice ergonomic risk factors found in the classroom the higher were the teachers' total scores on voice symptoms and VHI. Stress was the factor that correlated most strongly with voice symptoms. Poor indoor air quality increased the occurrence of laryngitis. Voice ergonomics were poor in the classrooms studied and voice ergonomic risk factors affected the voice. It is important to convey information on voice ergonomics to education administrators and those responsible for school planning and taking care of school buildings. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  3. The prevalence of voice disorders in 911 emergency telecommunicators.

    PubMed

    Johns-Fiedler, Heidi; van Mersbergen, Miriam

    2015-05-01

    Emergency 911 dispatchers or telecommunicators have been cited as occupational voice users who could be at risk for voice disorders. To test the theoretical assumption that the 911 emergency telecommunicators (911ETCs) are exposed to risk for voice disorders because of their heavy vocal load, this study assessed the prevalence of voice complaints in 911ETCs. A cross-sectional survey was sent to two large national organizations for 911ETCs with 71 complete responses providing information about voice health, voice complaints, and work load. Although 911ETCs have a higher rate of reported voice symptoms and score higher on the Voice Handicap Index-10 than the general public, they have a voice disorder diagnosis prevalence that mirrors the prevalence of the general population. The 911ETCs may be underserved in the voice community and would benefit from education on vocal health and treatments for voice complaints. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  4. Loud and angry: sound intensity modulates amygdala activation to angry voices in social anxiety disorder.

    PubMed

    Simon, Doerte; Becker, Michael; Mothes-Lasch, Martin; Miltner, Wolfgang H R; Straube, Thomas

    2017-03-01

    Angry expressions of both voices and faces represent disorder-relevant stimuli in social anxiety disorder (SAD). Although individuals with SAD show greater amygdala activation to angry faces, previous work has failed to find comparable effects for angry voices. Here, we investigated whether voice sound-intensity, a modulator of a voice's threat-relevance, affects brain responses to angry prosody in SAD. We used event-related functional magnetic resonance imaging to explore brain responses to voices varying in sound intensity and emotional prosody in SAD patients and healthy controls (HCs). Angry and neutral voices were presented either with normal or high sound amplitude, while participants had to decide upon the speaker's gender. Loud vs normal voices induced greater insula activation, and angry vs neutral prosody greater orbitofrontal cortex activation in SAD as compared with HC subjects. Importantly, an interaction of sound intensity, prosody and group was found in the insula and the amygdala. In particular, the amygdala showed greater activation to loud angry voices in SAD as compared with HC subjects. This finding demonstrates a modulating role of voice sound-intensity on amygdalar hyperresponsivity to angry prosody in SAD and suggests that abnormal processing of interpersonal threat signals in amygdala extends beyond facial expressions in SAD. © The Author (2016). Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.

  5. Loud and angry: sound intensity modulates amygdala activation to angry voices in social anxiety disorder

    PubMed Central

    Simon, Doerte; Becker, Michael; Mothes-Lasch, Martin; Miltner, Wolfgang H.R.

    2017-01-01

    Abstract Angry expressions of both voices and faces represent disorder-relevant stimuli in social anxiety disorder (SAD). Although individuals with SAD show greater amygdala activation to angry faces, previous work has failed to find comparable effects for angry voices. Here, we investigated whether voice sound-intensity, a modulator of a voice’s threat-relevance, affects brain responses to angry prosody in SAD. We used event-related functional magnetic resonance imaging to explore brain responses to voices varying in sound intensity and emotional prosody in SAD patients and healthy controls (HCs). Angry and neutral voices were presented either with normal or high sound amplitude, while participants had to decide upon the speaker’s gender. Loud vs normal voices induced greater insula activation, and angry vs neutral prosody greater orbitofrontal cortex activation in SAD as compared with HC subjects. Importantly, an interaction of sound intensity, prosody and group was found in the insula and the amygdala. In particular, the amygdala showed greater activation to loud angry voices in SAD as compared with HC subjects. This finding demonstrates a modulating role of voice sound-intensity on amygdalar hyperresponsivity to angry prosody in SAD and suggests that abnormal processing of interpersonal threat signals in amygdala extends beyond facial expressions in SAD. PMID:27651541

  6. Expectations and Experiences: The Voice of a First-Generation First-Year College Student and the Question of Student Persistence

    ERIC Educational Resources Information Center

    Stieha, Vicki

    2010-01-01

    This single case study takes a phenomenological approach using the voice centered analysis to analyze qualitative interview data so that the voice of this first-generation college student is brought forward. It is a poignant voice filled with conflicting emotional responses to the desire for college success, for family stability, for meaningful…

  7. When We Listen: Using Student Voices to Design Culturally Responsive and Just Schools

    ERIC Educational Resources Information Center

    Bunner, Teresa

    2017-01-01

    Teresa Bunner works as the high school literacy coordinating teacher for the Wake County (NC) Public School System in Raleigh. In this article she discusses the concept of Culturally Responsive teaching as a pedagogy that recognizes the importance of including students' cultural reference in all aspects of learning. She describes how she and her…

  8. Context, Contrast, and Tone of Voice in Auditory Sarcasm Perception.

    PubMed

    Voyer, Daniel; Thibodeau, Sophie-Hélène; Delong, Breanna J

    2016-02-01

    Four experiments were conducted to investigate the interplay between context and tone of voice in the perception of sarcasm. These experiments emphasized the role of contrast effects in sarcasm perception exclusively by means of auditory stimuli whereas most past research has relied on written material. In all experiments, a positive or negative computer-generated context spoken in a flat emotional tone was followed by a literally positive statement spoken in a sincere or sarcastic tone of voice. Participants indicated for each statement whether the intonation was sincere or sarcastic. In Experiment 1, a congruent context/tone of voice pairing (negative/sarcastic, positive/sincere) produced fast response times and proportions of sarcastic responses in the direction predicted by the tone of voice. Incongruent pairings produced mid-range proportions and slower response times. Experiment 2 introduced ambiguous contexts to determine whether a lower context/statements contrast would affect the proportion of sarcastic responses and response time. Results showed the expected findings for proportions (values between those obtained for congruent and incongruent pairings in the direction predicted by the tone of voice). However, response time failed to produce the predicted pattern, suggesting potential issues with the choice of stimuli. Experiments 3 and 4 extended the results of Experiments 1 and 2, respectively, to auditory stimuli based on written vignettes used in neuropsychological assessment. Results were exactly as predicted by contrast effects in both experiments. Taken together, the findings suggest that both context and tone influence how sarcasm is perceived while supporting the importance of contrast effects in sarcasm perception.

  9. Nurses using futuristic technology in today's healthcare setting.

    PubMed

    Wolf, Debra M; Kapadia, Amar; Kintzel, Jessie; Anton, Bonnie B

    2009-01-01

    Human computer interaction (HCI) equates nurses using voice assisted technology within a clinical setting to document patient care real time, retrieve patient information from care plans, and complete routine tasks. This is a reality currently utilized by clinicians today in acute and long term care settings. Voice assisted documentation provides hands & eyes free accurate documentation while enabling effective communication and task management. The speech technology increases the accuracy of documentation, while interfacing directly into the electronic health record (EHR). Using technology consisting of a light weight headset and small fist size wireless computer, verbal responses to easy to follow cues are converted into a database systems allowing staff to obtain individualized care status reports on demand. To further assist staff in their daily process, this innovative technology allows staff to send and receive pages as needed. This paper will discuss how leading edge and award winning technology is being integrated within the United States. Collaborative efforts between clinicians and analyst will be discussed reflecting the interactive design and build functionality. Features such as the system's voice responses and directed cues will be shared and how easily data can be documented, viewed and retrieved. Outcome data will be presented on how the technology impacted organization's quality outcomes, financial reimbursement, and employee's level of satisfaction.

  10. Neural basis of processing threatening voices in a crowded auditory world

    PubMed Central

    Mothes-Lasch, Martin; Becker, Michael P. I.; Miltner, Wolfgang H. R.

    2016-01-01

    In real world situations, we typically listen to voice prosody against a background crowded with auditory stimuli. Voices and background can both contain behaviorally relevant features and both can be selectively in the focus of attention. Adequate responses to threat-related voices under such conditions require that the brain unmixes reciprocally masked features depending on variable cognitive resources. It is unknown which brain systems instantiate the extraction of behaviorally relevant prosodic features under varying combinations of prosody valence, auditory background complexity and attentional focus. Here, we used event-related functional magnetic resonance imaging to investigate the effects of high background sound complexity and attentional focus on brain activation to angry and neutral prosody in humans. Results show that prosody effects in mid superior temporal cortex were gated by background complexity but not attention, while prosody effects in the amygdala and anterior superior temporal cortex were gated by attention but not background complexity, suggesting distinct emotional prosody processing limitations in different regions. Crucially, if attention was focused on the highly complex background, the differential processing of emotional prosody was prevented in all brain regions, suggesting that in a distracting, complex auditory world even threatening voices may go unnoticed. PMID:26884543

  11. An Examination of Selected Datacom Options for the Near-Term Implementation of Trajectory Based Operations

    NASA Technical Reports Server (NTRS)

    Johnson, Walter W.; Lachter, Joel B.; Battiste, Vernol; Lim, Veranika; Brandt, Summer L.; Koteskey, Robert W.; Dao, Arik-Quang V.; Ligda, Sarah V.; Wu, Shu-Chieh

    2011-01-01

    A primary feature of the Next Generation Air Transportation System (NextGen) is trajectory based operations (TBO). Under TBO, aircraft flight plans are known to computer systems on the ground that aid in scheduling and separation. The Future Air Navigation System (FANS) was developed to support TBO, but relatively few aircraft in the US are FANSequipped. Thus, any near-term implementation must provide TBO procedures for non-FANS aircraft. Previous research has explored controller clearances, but any implementation must also provide procedures for aircraft requests. The work presented here aims to surface issues surrounding TBO communication procedures for non-FANS aircraft and for aircraft requesting deviations around weather. Three types of communication were explored: Voice, FANS, and ACARS,(Aircraft Communications Addressing and Reporting System). ACARS and FANS are datacom systems that differ in that FANS allows uplinked flight plans to be loaded into the Flight Management System (FMS), while ACARS delivers flight plans as text that must be entered manually via the Control Display Unit (CDU). Sixteen pilots (eight two-person flight decks) and four controllers participated in 32 20-minute scenarios that required the flight decks to navigate through convective weather as they approached their top of descents (TODs). Findings: The rate of non-conformance was higher than anticipated, with aircraft off path more than 20% of the time. Controllers did not differentiate between the ACARS and FANS datacom, and were mixed in their preference for Voice vs. datacom (ACARS and FANS). Pilots uniformly preferred Voice to datacom, particularly ACARS. Much of their dislike appears to result from the slow response times in the datacom conditions. As a result, participants frequently resorted to voice communication. These results imply that, before implementing TBO in environments where pilots make weather deviation requests, further research is needed to develop communication procedures that integrate voice and datacom.

  12. ``The perceptual bases of speaker identity'' revisited

    NASA Astrophysics Data System (ADS)

    Voiers, William D.

    2003-10-01

    A series of experiments begun 40 years ago [W. D. Voiers, J. Acoust. Soc. Am. 36, 1065-1073 (1964)] was concerned with identifying the perceived voice traits (PVTs) on which human recognition of voices depends. It culminated with the development of a voice taxonomy based on 20 PVTs and a set of highly reliable rating scales for classifying voices with respect to those PVTs. The development of a perceptual voice taxonomy was motivated by the need for a practical method of evaluating speaker recognizability in voice communication systems. The Diagnostic Speaker Recognition Test (DSRT) evaluates the effects of systems on speaker recognizability as reflected in changes in the inter-listener reliability of voice ratings on the 20 PVTs. The DSRT thus provides a qualitative, as well as quantitative, evaluation of the effects of a system on speaker recognizability. A fringe benefit of this project is PVT rating data for a sample of 680 voices. [Work partially supported by USAFRL.

  13. Guidelines for Selecting Microphones for Human Voice Production Research

    ERIC Educational Resources Information Center

    Svec, Jan G.; Granqvist, Svante

    2010-01-01

    Purpose: This tutorial addresses fundamental characteristics of microphones (frequency response, frequency range, dynamic range, and directionality), which are important for accurate measurements of voice and speech. Method: Technical and voice literature was reviewed and analyzed. The following recommendations on desirable microphone…

  14. Initial Progress Toward Development of a Voice-Based Computer-Delivered Motivational Intervention for Heavy Drinking College Students: An Experimental Study.

    PubMed

    Kahler, Christopher W; Lechner, William J; MacGlashan, James; Wray, Tyler B; Littman, Michael L

    2017-06-28

    Computer-delivered interventions have been shown to be effective in reducing alcohol consumption in heavy drinking college students. However, these computer-delivered interventions rely on mouse, keyboard, or touchscreen responses for interactions between the users and the computer-delivered intervention. The principles of motivational interviewing suggest that in-person interventions may be effective, in part, because they encourage individuals to think through and speak aloud their motivations for changing a health behavior, which current computer-delivered interventions do not allow. The objective of this study was to take the initial steps toward development of a voice-based computer-delivered intervention that can ask open-ended questions and respond appropriately to users' verbal responses, more closely mirroring a human-delivered motivational intervention. We developed (1) a voice-based computer-delivered intervention that was run by a human controller and that allowed participants to speak their responses to scripted prompts delivered by speech generation software and (2) a text-based computer-delivered intervention that relied on the mouse, keyboard, and computer screen for all interactions. We randomized 60 heavy drinking college students to interact with the voice-based computer-delivered intervention and 30 to interact with the text-based computer-delivered intervention and compared their ratings of the systems as well as their motivation to change drinking and their drinking behavior at 1-month follow-up. Participants reported that the voice-based computer-delivered intervention engaged positively with them in the session and delivered content in a manner consistent with motivational interviewing principles. At 1-month follow-up, participants in the voice-based computer-delivered intervention condition reported significant decreases in quantity, frequency, and problems associated with drinking, and increased perceived importance of changing drinking behaviors. In comparison to the text-based computer-delivered intervention condition, those assigned to voice-based computer-delivered intervention reported significantly fewer alcohol-related problems at the 1-month follow-up (incident rate ratio 0.60, 95% CI 0.44-0.83, P=.002). The conditions did not differ significantly on perceived importance of changing drinking or on measures of drinking quantity and frequency of heavy drinking. Results indicate that it is feasible to construct a series of open-ended questions and a bank of responses and follow-up prompts that can be used in a future fully automated voice-based computer-delivered intervention that may mirror more closely human-delivered motivational interventions to reduce drinking. Such efforts will require using advanced speech recognition capabilities and machine-learning approaches to train a program to mirror the decisions made by human controllers in the voice-based computer-delivered intervention used in this study. In addition, future studies should examine enhancements that can increase the perceived warmth and empathy of voice-based computer-delivered intervention, possibly through greater personalization, improvements in the speech generation software, and embodying the computer-delivered intervention in a physical form. ©Christopher W Kahler, William J Lechner, James MacGlashan, Tyler B Wray, Michael L Littman. Originally published in JMIR Mental Health (http://mental.jmir.org), 28.06.2017.

  15. Natural asynchronies in audiovisual communication signals regulate neuronal multisensory interactions in voice-sensitive cortex.

    PubMed

    Perrodin, Catherine; Kayser, Christoph; Logothetis, Nikos K; Petkov, Christopher I

    2015-01-06

    When social animals communicate, the onset of informative content in one modality varies considerably relative to the other, such as when visual orofacial movements precede a vocalization. These naturally occurring asynchronies do not disrupt intelligibility or perceptual coherence. However, they occur on time scales where they likely affect integrative neuronal activity in ways that have remained unclear, especially for hierarchically downstream regions in which neurons exhibit temporally imprecise but highly selective responses to communication signals. To address this, we exploited naturally occurring face- and voice-onset asynchronies in primate vocalizations. Using these as stimuli we recorded cortical oscillations and neuronal spiking responses from functional MRI (fMRI)-localized voice-sensitive cortex in the anterior temporal lobe of macaques. We show that the onset of the visual face stimulus resets the phase of low-frequency oscillations, and that the face-voice asynchrony affects the prominence of two key types of neuronal multisensory responses: enhancement or suppression. Our findings show a three-way association between temporal delays in audiovisual communication signals, phase-resetting of ongoing oscillations, and the sign of multisensory responses. The results reveal how natural onset asynchronies in cross-sensory inputs regulate network oscillations and neuronal excitability in the voice-sensitive cortex of macaques, a suggested animal model for human voice areas. These findings also advance predictions on the impact of multisensory input on neuronal processes in face areas and other brain regions.

  16. Writing about rape: use of the passive voice and other distancing text features as an expression of perceived responsibility of the victim.

    PubMed

    Bohner, G

    2001-12-01

    The hypothesis that the passive voice is used to put the actor in the background and the acted-upon person in the focus of discourse is tested in the realm of sexual violence. German university students (N = 67) watched a silent video segment depicting a rape whose circumstances, depending on condition, could or could not be easily interpreted in terms of rape myths. Then they wrote down what they had seen, judged the responsibility of assailant and victim, and completed a rape-myth acceptance scale. Participants used the passive voice more frequently to describe the rape itself vs. other actions they had watched. When circumstances of the rape were easily interpretable in terms of rape myths, use of the passive voice correlated positively with rape-myth acceptance and perceived responsibility of the victim, and negatively with perceived responsibility of the assailant. The language of headlines that participants generated for their reports also reflected judgments of assailant and victim responsibility. Implications for the non-reactive assessment of responsibility attributions and directions for future research are discussed.

  17. Finding and Learning to Use the Singing Voice: A Manual for Teachers.

    ERIC Educational Resources Information Center

    Gould, A. Oren

    The child who is unable to reproduce a melody at a given pitch range can begin to "carry a tune" by learning to hear and control his singing voice and to match his voice with voices of other singers or with instruments. The "too low problem singer," the child with the most common difficulty, must learn to make successful song responses in his…

  18. Auditory and visual modulation of temporal lobe neurons in voice-sensitive and association cortices.

    PubMed

    Perrodin, Catherine; Kayser, Christoph; Logothetis, Nikos K; Petkov, Christopher I

    2014-02-12

    Effective interactions between conspecific individuals can depend upon the receiver forming a coherent multisensory representation of communication signals, such as merging voice and face content. Neuroimaging studies have identified face- or voice-sensitive areas (Belin et al., 2000; Petkov et al., 2008; Tsao et al., 2008), some of which have been proposed as candidate regions for face and voice integration (von Kriegstein et al., 2005). However, it was unclear how multisensory influences occur at the neuronal level within voice- or face-sensitive regions, especially compared with classically defined multisensory regions in temporal association cortex (Stein and Stanford, 2008). Here, we characterize auditory (voice) and visual (face) influences on neuronal responses in a right-hemisphere voice-sensitive region in the anterior supratemporal plane (STP) of Rhesus macaques. These results were compared with those in the neighboring superior temporal sulcus (STS). Within the STP, our results show auditory sensitivity to several vocal features, which was not evident in STS units. We also newly identify a functionally distinct neuronal subpopulation in the STP that appears to carry the area's sensitivity to voice identity related features. Audiovisual interactions were prominent in both the STP and STS. However, visual influences modulated the responses of STS neurons with greater specificity and were more often associated with congruent voice-face stimulus pairings than STP neurons. Together, the results reveal the neuronal processes subserving voice-sensitive fMRI activity patterns in primates, generate hypotheses for testing in the visual modality, and clarify the position of voice-sensitive areas within the unisensory and multisensory processing hierarchies.

  19. Auditory and Visual Modulation of Temporal Lobe Neurons in Voice-Sensitive and Association Cortices

    PubMed Central

    Perrodin, Catherine; Kayser, Christoph; Logothetis, Nikos K.

    2014-01-01

    Effective interactions between conspecific individuals can depend upon the receiver forming a coherent multisensory representation of communication signals, such as merging voice and face content. Neuroimaging studies have identified face- or voice-sensitive areas (Belin et al., 2000; Petkov et al., 2008; Tsao et al., 2008), some of which have been proposed as candidate regions for face and voice integration (von Kriegstein et al., 2005). However, it was unclear how multisensory influences occur at the neuronal level within voice- or face-sensitive regions, especially compared with classically defined multisensory regions in temporal association cortex (Stein and Stanford, 2008). Here, we characterize auditory (voice) and visual (face) influences on neuronal responses in a right-hemisphere voice-sensitive region in the anterior supratemporal plane (STP) of Rhesus macaques. These results were compared with those in the neighboring superior temporal sulcus (STS). Within the STP, our results show auditory sensitivity to several vocal features, which was not evident in STS units. We also newly identify a functionally distinct neuronal subpopulation in the STP that appears to carry the area's sensitivity to voice identity related features. Audiovisual interactions were prominent in both the STP and STS. However, visual influences modulated the responses of STS neurons with greater specificity and were more often associated with congruent voice-face stimulus pairings than STP neurons. Together, the results reveal the neuronal processes subserving voice-sensitive fMRI activity patterns in primates, generate hypotheses for testing in the visual modality, and clarify the position of voice-sensitive areas within the unisensory and multisensory processing hierarchies. PMID:24523543

  20. Dysphonia, Perceived Control, and Psychosocial Distress: A Qualitative Study.

    PubMed

    Misono, Stephanie; Haut, Caroline; Meredith, Liza; Frazier, Patricia A; Stockness, Ali; Michael, Deirdre D; Butcher, Lisa; Harwood, Eileen M

    2018-05-11

    The purpose of this qualitative study was to examine relationships between psychological factors, particularly perceived control, and voice symptoms in adults seeking treatment for a voice problem. Semistructured interviews of adult patients with a clinical diagnosis of muscle tension dysphonia were conducted and transcribed. Follow-up interviews were conducted as needed for further information or clarification. A multidisciplinary team analyzed interview content using inductive techniques. Common themes and subthemes were identified. A conceptual model was developed describing the association between voice symptoms, psychological factors, precipitants of ongoing voice symptoms, and perceived control. Thematic saturation was reached after 23 interviews. No participants reported a direct psychological cause for their voice problem, although half described significant life events preceding voice problem onset (eg, miscarriage and other health events, interpersonal conflicts, and family members' illnesses, injuries, and deaths). Participants described psychological influences on voice symptoms that led to rapid exacerbation of their voice symptoms. Participants described the helpfulness of speech therapy and sometimes also challenges of applying techniques in daily life. They also discussed personal coping strategies that included behavioral (eg, avoiding triggers and seeking social support) and psychological (eg, mind-body awareness and emotion regulation) components. Voice-related perceived control was associated with adaptive emotional and behavioral responses, which appeared to facilitate symptom improvement. In this qualitative pilot study, participant narratives suggested that psychological factors and emotions influence voice symptoms, facilitating development of a preliminary conceptual model of how adaptive and maladaptive responses develop and how they influence vocal function. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  1. Interactions between voice clinics and singing teachers: a report on the British Voice Association questionnaire to voice clinics in the UK.

    PubMed

    Davies, J; Anderson, S; Huchison, L; Stewart, G

    2007-01-01

    Singers with vocal problems are among patients who present at multidisciplinary voice clinics led by Ear Nose and Throat consultants and laryngologists or speech and language therapists. However, the development and care of the singing voice are also important responsibilities of singing teachers. We report here on the current extent and nature of interactions between voice clinics and singing teachers, based on data from a recent survey undertaken on behalf of the British Voice Association. A questionnaire was sent to all 103 voice clinics at National Health Service (NHS) hospitals in the UK. Responses were received and analysed from 42 currently active clinics. Eight (19%) clinics reported having a singing teacher as an active member of the team. They were all satisfied with the singing teacher's knowledge and expertise, which had been acquired by several different means. Of 32 clinics without a singing teacher regularly associated with the team, funding and difficulty of finding an appropriate singing voice expert (81% and 50%, respectively) were among the main reasons for their absence. There was an expressed requirement for more interaction between voice clinics and singing teachers, and 86% replied that they would find it useful to have a list of singing teachers in their area. On the matter of gaining expertise and training, 74% of the clinics replying would enable singing teachers to observe clinic sessions for experience and 21% were willing to assist in training them for clinic-associated work.

  2. Pilot study on the feasibility of a computerized speech recognition charting system.

    PubMed

    Feldman, C A; Stevens, D

    1990-08-01

    The objective of this study was to determine the feasibility of developing and using a voice recognition computerized charting system to record dental clinical examination data. More specifically, the study was designed to analyze the time and error differential between the traditional examiner/recorder method (ASSISTANT) and computerized voice recognition method (VOICE). DMFS examinations were performed twice on 20 patients using the traditional ASSISTANT and the VOICE charting system. A statistically significant difference was found when comparing the mean ASSISTANT time of 2.69 min to the VOICE time of 3.72 min (P less than 0.001). No statistically significant difference was found when comparing the mean ASSISTANT recording errors of 0.1 to VOICE recording errors of 0.6 (P = 0.059). 90% of the patients indicated they felt comfortable with the dentist talking to a computer and only 5% of the sample indicated they opposed VOICE. Results from this pilot study indicate that a charting system utilizing voice recognition technology could be considered a viable alternative to traditional examiner/recorder methods of clinical charting.

  3. Computer-automated dementia screening using a touch-tone telephone.

    PubMed

    Mundt, J C; Ferber, K L; Rizzo, M; Greist, J H

    2001-11-12

    This study investigated the sensitivity and specificity of a computer-automated telephone system to evaluate cognitive impairment in elderly callers to identify signs of early dementia. The Clinical Dementia Rating Scale was used to assess 155 subjects aged 56 to 93 years (n = 74, 27, 42, and 12, with a Clinical Dementia Rating Scale score of 0, 0.5, 1, and 2, respectively). These subjects performed a battery of tests administered by an interactive voice response system using standard Touch-Tone telephones. Seventy-four collateral informants also completed an interactive voice response version of the Symptoms of Dementia Screener. Sixteen cognitively impaired subjects were unable to complete the telephone call. Performances on 6 of 8 tasks were significantly influenced by Clinical Dementia Rating Scale status. The mean (SD) call length was 12 minutes 27 seconds (2 minutes 32 seconds). A subsample (n = 116) was analyzed using machine-learning methods, producing a scoring algorithm that combined performances across 4 tasks. Results indicated a potential sensitivity of 82.0% and specificity of 85.5%. The scoring model generalized to a validation subsample (n = 39), producing 85.0% sensitivity and 78.9% specificity. The kappa agreement between predicted and actual group membership was 0.64 (P<.001). Of the 16 subjects unable to complete the call, 11 provided sufficient information to permit us to classify them as impaired. Standard scoring of the interactive voice response-administered Symptoms of Dementia Screener (completed by informants) produced a screening sensitivity of 63.5% and 100% specificity. A lower criterion found a 90.4% sensitivity, without lowering specificity. Computer-automated telephone screening for early dementia using either informant or direct assessment is feasible. Such systems could provide wide-scale, cost-effective screening, education, and referral services to patients and caregivers.

  4. Early development of polyphonic sound encoding and the high voice superiority effect.

    PubMed

    Marie, Céline; Trainor, Laurel J

    2014-05-01

    Previous research suggests that when two streams of pitched tones are presented simultaneously, adults process each stream in a separate memory trace, as reflected by mismatch negativity (MMN), a component of the event-related potential (ERP). Furthermore, a superior encoding of the higher tone or voice in polyphonic sounds has been found for 7-month-old infants and both musician and non-musician adults in terms of a larger amplitude MMN in response to pitch deviant stimuli in the higher than the lower voice. These results, in conjunction with modeling work, suggest that the high voice superiority effect might originate in characteristics of the peripheral auditory system. If this is the case, the high voice superiority effect should be present in infants younger than 7 months. In the present study we tested 3-month-old infants as there is no evidence at this age of perceptual narrowing or specialization of musical processing according to the pitch or rhythmic structure of music experienced in the infant׳s environment. We presented two simultaneous streams of tones (high and low) with 50% of trials modified by 1 semitone (up or down), either on the higher or the lower tone, leaving 50% standard trials. Results indicate that like the 7-month-olds, 3-month-old infants process each tone in a separate memory trace and show greater saliency for the higher tone. Although MMN was smaller and later in both voices for the group of sixteen 3-month-olds compared to the group of sixteen 7-month-olds, the size of the difference in MMN for the high compared to low voice was similar across ages. These results support the hypothesis of an innate peripheral origin of the high voice superiority effect. Copyright © 2014 Elsevier Ltd. All rights reserved.

  5. Voice-on-Target: A New Approach to Tactical Networking and Unmanned Systems Control via the Voice Interface to the SA Environment

    DTIC Science & Technology

    2009-06-01

    Blackberry handheld) device. After each voice command activation, the medic provided voice comments to be recorded in Observer Notepad over Voice...vial (up-right corner of picture) upon voice activation from the medic’s Blackberry handheld. The NPS UAS which was controlled by voice commands...Voice Portal using a standard Blackberry handheld with a head set. The results demonstrated sufficient accuracy for controlling the tactical sensor

  6. The persuasiveness of synthetic speech versus human speech.

    PubMed

    Stern, S E; Mullennix, J W; Dyson, C; Wilson, S J

    1999-12-01

    Is computer-synthesized speech as persuasive as the human voice when presenting an argument? After completing an attitude pretest, 193 participants were randomly assigned to listen to a persuasive appeal under three conditions: a high-quality synthesized speech system (DECtalk Express), a low-quality synthesized speech system (Monologue), and a tape recording of a human voice. Following the appeal, participants completed a posttest attitude survey and a series of questionnaires designed to assess perceptions of speech qualities, perceptions of the speaker, and perceptions of the message. The human voice was generally perceived more favorably than the computer-synthesized voice, and the speaker was perceived more favorably when the voice was a human voice than when it was computer synthesized. There was, however, no evidence that computerized speech, as compared with the human voice, affected persuasion or perceptions of the message. Actual or potential applications of this research include issues that should be considered when designing synthetic speech systems.

  7. Natural asynchronies in audiovisual communication signals regulate neuronal multisensory interactions in voice-sensitive cortex

    PubMed Central

    Perrodin, Catherine; Kayser, Christoph; Logothetis, Nikos K.; Petkov, Christopher I.

    2015-01-01

    When social animals communicate, the onset of informative content in one modality varies considerably relative to the other, such as when visual orofacial movements precede a vocalization. These naturally occurring asynchronies do not disrupt intelligibility or perceptual coherence. However, they occur on time scales where they likely affect integrative neuronal activity in ways that have remained unclear, especially for hierarchically downstream regions in which neurons exhibit temporally imprecise but highly selective responses to communication signals. To address this, we exploited naturally occurring face- and voice-onset asynchronies in primate vocalizations. Using these as stimuli we recorded cortical oscillations and neuronal spiking responses from functional MRI (fMRI)-localized voice-sensitive cortex in the anterior temporal lobe of macaques. We show that the onset of the visual face stimulus resets the phase of low-frequency oscillations, and that the face–voice asynchrony affects the prominence of two key types of neuronal multisensory responses: enhancement or suppression. Our findings show a three-way association between temporal delays in audiovisual communication signals, phase-resetting of ongoing oscillations, and the sign of multisensory responses. The results reveal how natural onset asynchronies in cross-sensory inputs regulate network oscillations and neuronal excitability in the voice-sensitive cortex of macaques, a suggested animal model for human voice areas. These findings also advance predictions on the impact of multisensory input on neuronal processes in face areas and other brain regions. PMID:25535356

  8. Memory strength and specificity revealed by pupillometry

    PubMed Central

    Papesh, Megan H.; Goldinger, Stephen D.; Hout, Michael C.

    2011-01-01

    Voice-specificity effects in recognition memory were investigated using both behavioral data and pupillometry. Volunteers initially heard spoken words and nonwords in two voices; they later provided confidence-based old/new classifications to items presented in their original voices, changed (but familiar) voices, or entirely new voices. Recognition was more accurate for old-voice items, replicating prior research. Pupillometry was used to gauge cognitive demand during both encoding and testing: Enlarged pupils revealed that participants devoted greater effort to encoding items that were subsequently recognized. Further, pupil responses were sensitive to the cue match between encoding and retrieval voices, as well as memory strength. Strong memories, and those with the closest encoding-retrieval voice matches, resulted in the highest peak pupil diameters. The results are discussed with respect to episodic memory models and Whittlesea’s (1997) SCAPE framework for recognition memory. PMID:22019480

  9. Call progress time measurement in IP telephony

    NASA Astrophysics Data System (ADS)

    Khasnabish, Bhumip

    1999-11-01

    Usually a voice call is established through multiple stages in IP telephony. In the first stage, a phone number is dialed to reach a near-end or call-originating IP-telephony gateway. The next stages involve user identification through delivering an m-digit user-id to the authentication and/or billing server, and then user authentication by using an n- digit PIN. After that, the caller is allowed (last stage dial tone is provided) to dial a destination phone number provided that authentication is successful. In this paper, we present a very flexible method for measuring call progress time in IP telephony. The proposed technique can be used to measure the system response time at every stage. It is flexible, so that it can be easily modified to include new `tone' or a set of tones, or `voice begin' can be used in every stage to detect the system's response. The proposed method has been implemented using scripts written in Hammer visual basic language for testing with a few commercially available IP telephony gateways.

  10. Swinging at a cocktail party: voice familiarity aids speech perception in the presence of a competing voice.

    PubMed

    Johnsrude, Ingrid S; Mackey, Allison; Hakyemez, Hélène; Alexander, Elizabeth; Trang, Heather P; Carlyon, Robert P

    2013-10-01

    People often have to listen to someone speak in the presence of competing voices. Much is known about the acoustic cues used to overcome this challenge, but almost nothing is known about the utility of cues derived from experience with particular voices--cues that may be particularly important for older people and others with impaired hearing. Here, we use a version of the coordinate-response-measure procedure to show that people can exploit knowledge of a highly familiar voice (their spouse's) not only to track it better in the presence of an interfering stranger's voice, but also, crucially, to ignore it so as to comprehend a stranger's voice more effectively. Although performance declines with increasing age when the target voice is novel, there is no decline when the target voice belongs to the listener's spouse. This finding indicates that older listeners can exploit their familiarity with a speaker's voice to mitigate the effects of sensory and cognitive decline.

  11. Telemedicine to promote patient safety: Use of phone-based interactive voice response system (IVRS) to reduce adverse safety events in predialysis CKD

    PubMed Central

    Weiner, Shoshana; Fink, Jeffery C.

    2017-01-01

    Chronic kidney disease (CKD) patients have several features conferring upon them a high risk of adverse safety events, which are defined as incidents with unintended harm related to processes of care or medications. These characteristics include impaired renal function, polypharmacy, and frequent health system encounters. The consequences of such events in CKD can include new or prolonged hospitalization, accelerated renal function loss, acute kidney injury, end-stage renal disease and death. Health information technology administered via telemedicine presents opportunities for CKD patients to remotely communicate safety-related findings to providers for the purpose of improving their care. However, many CKD patients have limitations which hinder their use of telemedicine and access to the broad capabilities of health information technology. In this review we summarize previous assessments of the pre-dialysis CKD populations’ proficiency in using telemedicine modalities and describe the use of interactive voice-response system (IVRS) to gauge the safety phenotype of the CKD patient. We discuss the potential for expanded IVRS use in CKD to address the safety threats inherent to this population. PMID:28224940

  12. Near-term fetal response to maternal spoken voice

    PubMed Central

    Voegtline, Kristin M.; Costigan, Kathleen A.; Pater, Heather A.; DiPietro, Janet A.

    2013-01-01

    Knowledge about prenatal learning has been largely predicated on the observation that newborns appear to recognize the maternal voice. Few studies have examined the process underlying this phenomenon; that is, whether and how the fetus responds to maternal voice in situ. Fetal heart rate and motor activity were recorded at 36 weeks gestation (n = 69) while pregnant women read aloud from a neutral passage. Compared to a baseline period, fetuses responded with a decrease in motor activity in the 10-seconds following onset of maternal speech and a trend level decelerative heart rate response, consistent with an orienting response. Subsequent analyses revealed that the fetal response was modified by both maternal and fetal factors. Fetuses of women who were previously awake and talking (n = 40) showed an orienting response to onset of maternal reading aloud, while fetuses of mothers who had previously been resting and silent (n = 29) responded with elevated heart rate and increased movement. The magnitude of the fetal response was further dependent on baseline fetal heart rate variability such that largest response was demonstrated by fetuses with low variability of mothers who were previously resting and silent. Results indicate that fetal responsivity is affected by both maternal and fetal state and have implications for understanding fetal learning of the maternal voice under naturalistic conditions. PMID:23748167

  13. Measuring the intuitive response of users when faced with different interactive paradigms to control a gastroenterology CAD system.

    PubMed

    Abrantes, D; Gomes, P; Pereira, D; Coimbra, M

    2016-08-01

    The gastroenterology specialty could benefit from the introduction of Computer Assisted Decision (CAD) systems, since gastric cancer is a serious concern in which an accurate and early diagnosis usually leads to a good prognosis. Still, the way doctors interact with these systems is very important because it will often determine its embracement or rejection, as any gains in productivity will frequently hinge on how comfortable they are with it. Using other types of interaction paradigms such as voice and motion control, is important in a way that typical inputs such as keyboard and mouse are sometimes not the best choice for certain clinical scenarios. In order to ascertain how a doctor could control a hypothetical CAD system during a gastroenterology exam, we measured the natural response of users when faced with three different task requests, using three types of interaction paradigms: voice, gesture and endoscope. Results fit in what was expected, with gesture control being the most intuitive to use, and the endoscope being on the other edge. All the technologies are mature enough to cope with the response concepts the participants gave us. However, when having into account the scenario context, better natural response scores may not always be the best choice for implementation. That way, simplification or reduction of tasks, along with a well tought-out interface, or even mixing more oriented paradigms for particular requests, could allow for better system control with fewer inconveniences for the user.

  14. On the definition and interpretation of voice selective activation in the temporal cortex

    PubMed Central

    Bethmann, Anja; Brechmann, André

    2014-01-01

    Regions along the superior temporal sulci and in the anterior temporal lobes have been found to be involved in voice processing. It has even been argued that parts of the temporal cortices serve as voice-selective areas. Yet, evidence for voice-selective activation in the strict sense is still missing. The current fMRI study aimed at assessing the degree of voice-specific processing in different parts of the superior and middle temporal cortices. To this end, voices of famous persons were contrasted with widely different categories, which were sounds of animals and musical instruments. The argumentation was that only brain regions with statistically proven absence of activation by the control stimuli may be considered as candidates for voice-selective areas. Neural activity was found to be stronger in response to human voices in all analyzed parts of the temporal lobes except for the middle and posterior STG. More importantly, the activation differences between voices and the other environmental sounds increased continuously from the mid-posterior STG to the anterior MTG. Here, only voices but not the control stimuli excited an increase of the BOLD response above a resting baseline level. The findings are discussed with reference to the function of the anterior temporal lobes in person recognition and the general question on how to define selectivity of brain regions for a specific class of stimuli or tasks. In addition, our results corroborate recent assumptions about the hierarchical organization of auditory processing building on a processing stream from the primary auditory cortices to anterior portions of the temporal lobes. PMID:25071527

  15. A Voice-Based E-Examination Framework for Visually Impaired Students in Open and Distance Learning

    ERIC Educational Resources Information Center

    Azeta, Ambrose A.; Inam, Itorobong A.; Daramola, Olawande

    2018-01-01

    Voice-based systems allow users access to information on the internet over a voice interface. Prior studies on Open and Distance Learning (ODL) e-examination systems that make use of voice interface do not sufficiently exhibit intelligent form of assessment, which diminishes the rigor of examination. The objective of this paper is to improve on…

  16. An Analysis of Content Delivery Systems Using Speaking Voice, Speaking with Repetition Voice, Chanting Voice, and Singing Voice.

    ERIC Educational Resources Information Center

    Foster, Karen R.; Kersh, Mildred E.; Masztal, Nancy B.

    This study investigated the way kindergarten classroom teachers delivered information to students to see if it affected the amount of information students could remember about the solar system. The study also examined whether this difference would be related to the degree of musical aptitude possessed by each student. The students were pretested…

  17. Design of digital voice storage and playback system

    NASA Astrophysics Data System (ADS)

    Tang, Chao

    2018-03-01

    Based on STC89C52 chip, this paper presents a single chip microcomputer minimum system, which is used to realize the logic control of digital speech storage and playback system. Compared with the traditional tape voice recording system, the system has advantages of small size, low power consumption, The effective solution of traditional voice recording system is limited in the use of electronic and information processing.

  18. Development from childhood to adulthood increases morphological and functional inter-individual variability in the right superior temporal cortex.

    PubMed

    Bonte, Milene; Frost, Martin A; Rutten, Sanne; Ley, Anke; Formisano, Elia; Goebel, Rainer

    2013-12-01

    We study the developmental trajectory of morphology and function of the superior temporal cortex (STC) in children (8-9 years), adolescents (14-15 years) and young adults. We analyze cortical surface landmarks and functional MRI (fMRI) responses to voices, other natural categories and tones and examine how hemispheric asymmetry and inter-subject variability change across age. Our results show stable morphological asymmetries across age groups, including a larger left planum temporale and a deeper right superior temporal sulcus. fMRI analyses show that a rightward lateralization for voice-selective responses is present in all groups but decreases with age. Furthermore, STC responses to voices change from being less selective and more spatially diffuse in children to highly selective and focal in adults. Interestingly, the analysis of morphological landmarks reveals that inter-subject variability increases during development in the right--but not in the left--STC. Similarly, inter-subject variability of cortically-realigned functional responses to voices, other categories and tones increases with age in the right STC. Our findings reveal asymmetric developmental changes in brain regions crucial for auditory and voice perception. The age-related increase of inter-subject variability in right STC suggests that anatomy and function of this region are shaped by unique individual developmental experiences. © 2013.

  19. Unlocking Elementary Students' Perspectives of Leadership

    ERIC Educational Resources Information Center

    Damiani, Jonathan

    2014-01-01

    This study examines whether and how principals take their lead from students, and use student voice, to create more responsive schools, and more responsible models of leadership. I consider issues of student agency and voice within four very different elementary school settings. Further, I consider the challenges students face, and the ways…

  20. Andreas Vesalius' 500th Anniversary: Initial Integral Understanding of Voice Production.

    PubMed

    Brinkman, Romy J; Hage, J Joris

    2017-01-01

    Voice production relies on the integrated functioning of a three-part system: respiration, phonation and resonance, and articulation. To commemorate the 500th anniversary of the great anatomist Andreas Vesalius (1515-1564), we report on his understanding of this integral system. The text of Vesalius' masterpiece De Humani Corporis Fabrica Libri Septum and an eyewitness report of the public dissection of three corpses by Vesalius in Bologna, Italy, in 1540, were searched for references to the voice-producing anatomical structures and their function. We clustered the traced, separate parts for the first time. We found that Vesalius recognized the importance for voice production of many details of the respiratory system, the voice box, and various structures of resonance and articulation. He stressed that voice production was a cerebral function and extensively recorded the innervation of the voice-producing organs by the cranial nerves. Vesalius was the first to publicly record the concept of voice production as an integrated and cerebrally directed function of respiration, phonation and resonance, and articulation. In doing so nearly 500 years ago, he laid a firm basis for the understanding of the physiology of voice production and speech and its management as we know it today. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  1. Identity Orientation, Voice, and Judgments of Procedural Justice during Late Adolescence

    ERIC Educational Resources Information Center

    Fondacaro, Mark R.; Brank, Eve M.; Stuart, Jennifer; Villanueva-Abraham, Sara; Luescher, Jennifer; McNatt, Penny S.

    2006-01-01

    This study focused on the relationship between voice and judgments of procedural justice in a sample of older adolescents and examined potential moderating and mediating influences of identity orientation (personal, social, and collective) and negative emotional response. Participants read 1 of 2 different family conflict scenarios (voice and no…

  2. The Development of Apt Citizenship Education through Listening to Young People's Voices

    ERIC Educational Resources Information Center

    Warwick, Paul

    2008-01-01

    Citizenship Education (CE) and the young people's voice agenda are both enjoying increasing popularity within England at the present time. Clear connections exist between the two, with CE placing an emphasis upon participation and responsible action and the young people's voice agenda advocating democratic procedures for involving young people in…

  3. Scientific bases of human-machine communication by voice.

    PubMed Central

    Schafer, R W

    1995-01-01

    The scientific bases for human-machine communication by voice are in the fields of psychology, linguistics, acoustics, signal processing, computer science, and integrated circuit technology. The purpose of this paper is to highlight the basic scientific and technological issues in human-machine communication by voice and to point out areas of future research opportunity. The discussion is organized around the following major issues in implementing human-machine voice communication systems: (i) hardware/software implementation of the system, (ii) speech synthesis for voice output, (iii) speech recognition and understanding for voice input, and (iv) usability factors related to how humans interact with machines. PMID:7479802

  4. Teacher response to ambulatory monitoring of voice.

    PubMed

    Hunter, Eric J

    2012-10-01

    Voice accumulation and dosimetry devices are used for unobtrusive monitoring of voice use. While numerous studies have used these devices to examine how individuals use their voices, little attention has been paid to how subjects respond to them. Therefore, the purpose of this short communication is to begin to explore two questions: 1) How do voice monitoring devices affect daily communication? and 2) How do participants feel about the physical design and function of these types of voice monitoring devices? One key finding is that most of the subjects remain aware of the dosimeter while wearing it, which may impact the data collected. Further, most subjects have difficulty with the accelerometer and/or the data storage device.

  5. Practical applications of interactive voice technologies: Some accomplishments and prospects

    NASA Technical Reports Server (NTRS)

    Grady, Michael W.; Hicklin, M. B.; Porter, J. E.

    1977-01-01

    A technology assessment of the application of computers and electronics to complex systems is presented. Three existing systems which utilize voice technology (speech recognition and speech generation) are described. Future directions in voice technology are also described.

  6. Utilization of Internet Protocol-Based Voice Systems in Remote Payload Operations

    NASA Technical Reports Server (NTRS)

    Best, Susan; Nichols, Kelvin; Bradford, Robert

    2003-01-01

    This viewgraph presentation provides an overview of a proposed voice communication system for use in remote payload operations performed on the International Space Station. The system, Internet Voice Distribution System (IVoDS), would make use of existing Internet protocols, and offer a number of advantages over the system currently in use. Topics covered include: system description and operation, system software and hardware, system architecture, project status, and technology transfer applications.

  7. Native voice, self-concept and the moral case for personalized voice technology.

    PubMed

    Nathanson, Esther

    2017-01-01

    Purpose (1) To explore the role of native voice and effects of voice loss on self-concept and identity, and survey the state of assistive voice technology; (2) to establish the moral case for developing personalized voice technology. Methods This narrative review examines published literature on the human significance of voice, the impact of voice loss on self-concept and identity, and the strengths and limitations of current voice technology. Based on the impact of voice loss on self and identity, and voice technology limitations, the moral case for personalized voice technology is developed. Results Given the richness of information conveyed by voice, loss of voice constrains expression of the self, but the full impact is poorly understood. Augmentative and alternative communication (AAC) devices facilitate communication but, despite advances in this field, voice output cannot yet express the unique nuances of individual voice. The ethical principles of autonomy, beneficence and equality of opportunity establish the moral responsibility to invest in accessible, cost-effective, personalized voice technology. Conclusions Although further research is needed to elucidate the full effects of voice loss on self-concept, identity and social functioning, current understanding of the profoundly negative impact of voice loss establishes the moral case for developing personalized voice technology. Implications for Rehabilitation Rehabilitation of voice-disordered patients should facilitate self-expression, interpersonal connectedness and social/occupational participation. Proactive questioning about the psychological and social experiences of patients with voice loss is a valuable entry point for rehabilitation planning. Personalized voice technology would enhance sense of self, communicative participation and autonomy and promote shared healthcare decision-making. Further research is needed to identify the best strategies to preserve and strengthen identity and sense of self.

  8. Cerebral Processing of Voice Gender Studied Using a Continuous Carryover fMRI Design

    PubMed Central

    Pernet, Cyril; Latinus, Marianne; Crabbe, Frances; Belin, Pascal

    2013-01-01

    Normal listeners effortlessly determine a person's gender by voice, but the cerebral mechanisms underlying this ability remain unclear. Here, we demonstrate 2 stages of cerebral processing during voice gender categorization. Using voice morphing along with an adaptation-optimized functional magnetic resonance imaging design, we found that secondary auditory cortex including the anterior part of the temporal voice areas in the right hemisphere responded primarily to acoustical distance with the previously heard stimulus. In contrast, a network of bilateral regions involving inferior prefrontal and anterior and posterior cingulate cortex reflected perceived stimulus ambiguity. These findings suggest that voice gender recognition involves neuronal populations along the auditory ventral stream responsible for auditory feature extraction, functioning in pair with the prefrontal cortex in voice gender perception. PMID:22490550

  9. Eye Movements Reveal Fast, Voice-Specific Priming

    PubMed Central

    Papesh, Megan H.; Goldinger, Stephen D.; Hout, Michael C.

    2015-01-01

    In spoken word perception, voice specificity effects are well-documented: When people hear repeated words in some task, performance is generally better when repeated items are presented in their originally heard voices, relative to changed voices. A key theoretical question about voice specificity effects concerns their time-course: Some studies suggest that episodic traces exert their influence late in lexical processing (the time-course hypothesis; McLennan & Luce, 2005), whereas others suggest that episodic traces influence immediate, online processing. We report two eye-tracking studies investigating the time-course of voice-specific priming within and across cognitive tasks. In Experiment 1, participants performed modified lexical decision or semantic classification to words spoken by four speakers. The tasks required participants to click a red “×” or a blue “+” located randomly within separate visual half-fields, necessitating trial-by-trial visual search with consistent half-field response mapping. After a break, participants completed a second block with new and repeated items, half spoken in changed voices. Voice effects were robust very early, appearing in saccade initiation times. Experiment 2 replicated this pattern while changing tasks across blocks, ruling out a response priming account. In the General Discussion, we address the time-course hypothesis, focusing on the challenge it presents for empirical disconfirmation, and highlighting the broad importance of indexical effects, beyond studies of priming. PMID:26726911

  10. 5 CFR 1690.1 - Definitions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ...Line means the automated voice response system by which TSP participants may, among other things... Internet location maintained by the Board, which contains information about the TSP and by which TSP participants may, among other things, access their accounts by computer. The TSP Web site address is www.tsp...

  11. Can a computer-generated voice be sincere? A case study combining music and synthetic speech.

    PubMed

    Barker, Paul; Newell, Christopher; Newell, George

    2013-10-01

    This article explores enhancing sincerity, honesty, or truthfulness in computer-generated synthetic speech by accompanying it with music. Sincerity is important if we are to respond positively to any voice, whether human or artificial. What is sincerity in the artificial disembodied voice? Studies in musical expression and performance may illuminate aspects of the 'musically spoken' or sung voice in rendering deeper levels of expression that may include sincerity. We consider one response to this notion in an especially composed melodrama (music accompanying a (synthetic) spoken voice) designed to convey sincerity.

  12. NWR (National Weather Service) voice synthesis project, phase 1

    NASA Astrophysics Data System (ADS)

    Sampson, G. W.

    1986-01-01

    The purpose of the NOAA Weather Radio (NWR) Voice Synthesis Project is to provide a demonstration of the current voice synthesis technology. Phase 1 of this project is presented, providing a complete automation of an hourly surface aviation observation for broadcast over NWR. In examining the products currently available on the market, the decision was made that synthetic voice technology does not have the high quality speech required for broadcast over the NWR. Therefore the system presented uses the phrase concatenation type of technology for a very high quality, versatile, voice synthesis system.

  13. Identification and human condition analysis based on the human voice analysis

    NASA Astrophysics Data System (ADS)

    Mieshkov, Oleksandr Yu.; Novikov, Oleksandr O.; Novikov, Vsevolod O.; Fainzilberg, Leonid S.; Kotyra, Andrzej; Smailova, Saule; Kozbekova, Ainur; Imanbek, Baglan

    2017-08-01

    The paper presents a two-stage biotechnical system for human condition analysis that is based on analysis of human voice signal. At the initial stage, the voice signal is pre-processed and its characteristics in time domain are determined. At the first stage, the developed system is capable of identifying the person in the database on the basis of the extracted characteristics. At the second stage, the model of a human voice is built on the basis of the real voice signals after clustering the whole database.

  14. STS-41 Voice Command System Flight Experiment Report

    NASA Technical Reports Server (NTRS)

    Salazar, George A.

    1981-01-01

    This report presents the results of the Voice Command System (VCS) flight experiment on the five-day STS-41 mission. Two mission specialists,Bill Shepherd and Bruce Melnick, used the speaker-dependent system to evaluate the operational effectiveness of using voice to control a spacecraft system. In addition, data was gathered to analyze the effects of microgravity on speech recognition performance.

  15. Feasibility of automated speech sample collection with stuttering children using interactive voice response (IVR) technology.

    PubMed

    Vogel, Adam P; Block, Susan; Kefalianos, Elaina; Onslow, Mark; Eadie, Patricia; Barth, Ben; Conway, Laura; Mundt, James C; Reilly, Sheena

    2015-04-01

    To investigate the feasibility of adopting automated interactive voice response (IVR) technology for remotely capturing standardized speech samples from stuttering children. Participants were 10 6-year-old stuttering children. Their parents called a toll-free number from their homes and were prompted to elicit speech from their children using a standard protocol involving conversation, picture description and games. The automated IVR system was implemented using an off-the-shelf telephony software program and delivered by a standard desktop computer. The software infrastructure utilizes voice over internet protocol. Speech samples were automatically recorded during the calls. Video recordings were simultaneously acquired in the home at the time of the call to evaluate the fidelity of the telephone collected samples. Key outcome measures included syllables spoken, percentage of syllables stuttered and an overall rating of stuttering severity using a 10-point scale. Data revealed a high level of relative reliability in terms of intra-class correlation between the video and telephone acquired samples on all outcome measures during the conversation task. Findings were less consistent for speech samples during picture description and games. Results suggest that IVR technology can be used successfully to automate remote capture of child speech samples.

  16. Study to determine potential flight applications and human factors design guidelines for voice recognition and synthesis systems

    NASA Astrophysics Data System (ADS)

    White, R. W.; Parks, D. L.

    1985-07-01

    A study was conducted to determine potential commercial aircraft flight deck applications and implementation guidelines for voice recognition and synthesis. At first, a survey of voice recognition and synthesis technology was undertaken to develop a working knowledge base. Then, numerous potential aircraft and simulator flight deck voice applications were identified and each proposed application was rated on a number of criteria in order to achieve an overall payoff rating. The potential voice recognition applications fell into five general categories: programming, interrogation, data entry, switch and mode selection, and continuous/time-critical action control. The ratings of the first three categories showed the most promise of being beneficial to flight deck operations. Possible applications of voice synthesis systems were categorized as automatic or pilot selectable and many were rated as being potentially beneficial. In addition, voice system implementation guidelines and pertinent performance criteria are proposed. Finally, the findings of this study are compared with those made in a recent NASA study of a 1995 transport concept.

  17. Study to determine potential flight applications and human factors design guidelines for voice recognition and synthesis systems

    NASA Technical Reports Server (NTRS)

    White, R. W.; Parks, D. L.

    1985-01-01

    A study was conducted to determine potential commercial aircraft flight deck applications and implementation guidelines for voice recognition and synthesis. At first, a survey of voice recognition and synthesis technology was undertaken to develop a working knowledge base. Then, numerous potential aircraft and simulator flight deck voice applications were identified and each proposed application was rated on a number of criteria in order to achieve an overall payoff rating. The potential voice recognition applications fell into five general categories: programming, interrogation, data entry, switch and mode selection, and continuous/time-critical action control. The ratings of the first three categories showed the most promise of being beneficial to flight deck operations. Possible applications of voice synthesis systems were categorized as automatic or pilot selectable and many were rated as being potentially beneficial. In addition, voice system implementation guidelines and pertinent performance criteria are proposed. Finally, the findings of this study are compared with those made in a recent NASA study of a 1995 transport concept.

  18. ‘Inner voices’: the cerebral representation of emotional voice cues described in literary texts

    PubMed Central

    Kreifelts, Benjamin; Gößling-Arnold, Christina; Wertheimer, Jürgen; Wildgruber, Dirk

    2014-01-01

    While non-verbal affective voice cues are generally recognized as a crucial behavioral guide in any day-to-day conversation their role as a powerful source of information may extend well beyond close-up personal interactions and include other modes of communication such as written discourse or literature as well. Building on the assumption that similarities between the different ‘modes’ of voice cues may not only be limited to their functional role but may also include cerebral mechanisms engaged in the decoding process, the present functional magnetic resonance imaging study aimed at exploring brain responses associated with processing emotional voice signals described in literary texts. Emphasis was placed on evaluating ‘voice’ sensitive as well as task- and emotion-related modulations of brain activation frequently associated with the decoding of acoustic vocal cues. Obtained findings suggest that several similarities emerge with respect to the perception of acoustic voice signals: results identify the superior temporal, lateral and medial frontal cortex as well as the posterior cingulate cortex and cerebellum to contribute to the decoding process, with similarities to acoustic voice perception reflected in a ‘voice’-cue preference of temporal voice areas as well as an emotion-related modulation of the medial frontal cortex and a task-modulated response of the lateral frontal cortex. PMID:24396008

  19. Control of voice fundamental frequency in speaking versus singing

    NASA Astrophysics Data System (ADS)

    Natke, Ulrich; Donath, Thomas M.; Kalveram, Karl Th.

    2003-03-01

    In order to investigate control of voice fundamental frequency (F0) in speaking and singing, 24 adults had to utter the nonsense word ['ta:tatas] repeatedly, while in selected trials their auditory feedback was frequency-shifted by 100 cents downwards. In the speaking condition the target speech rate and prosodic pattern were indicated by a rhythmic sequence made of white noise. In the singing condition the sequence consisted of piano notes, and subjects were instructed to match the pitch of the notes. In both conditions a response in voice F0 begins with a latency of about 150 ms. As predicted, response magnitude is greater in the singing condition (66 cents) than in the speaking condition (47 cents). Furthermore the singing condition seems to prolong the after-effect which is a continuation of the response in trials after the frequency shift. In the singing condition, response magnitude and the ability to match the target F0 correlate significantly. Results support the view that in speaking voice F0 is monitored mainly supra-segmentally and controlled less tightly than in singing.

  20. Control of voice fundamental frequency in speaking versus singing.

    PubMed

    Natke, Ulrich; Donath, Thomas M; Kalveram, Karl Th

    2003-03-01

    In order to investigate control of voice fundamental frequency (F0) in speaking and singing, 24 adults had to utter the nonsense word ['ta:tatas] repeatedly, while in selected trials their auditory feedback was frequency-shifted by 100 cents downwards. In the speaking condition the target speech rate and prosodic pattern were indicated by a rhythmic sequence made of white noise. In the singing condition the sequence consisted of piano notes, and subjects were instructed to match the pitch of the notes. In both conditions a response in voice F0 begins with a latency of about 150 ms. As predicted, response magnitude is greater in the singing condition (66 cents) than in the speaking condition (47 cents). Furthermore the singing condition seems to prolong the after-effect which is a continuation of the response in trials after the frequency shift. In the singing condition, response magnitude and the ability to match the target F0 correlate significantly. Results support the view that in speaking voice F0 is monitored mainly supra-segmentally and controlled less tightly than in singing.

  1. Clinical Features of Psychogenic Voice Disorder and the Efficiency of Voice Therapy and Psychological Evaluation.

    PubMed

    Tezcaner, Zahide Çiler; Gökmen, Muhammed Fatih; Yıldırım, Sibel; Dursun, Gürsel

    2017-11-06

    The aim of this study was to define the clinical features of psychogenic voice disorder (PVD) and explore the treatment efficiency of voice therapy and psychological evaluation. Fifty-eight patients who received treatment following the PVD diagnosis and had no organic or other functional voice disorders were assessed retrospectively based on laryngoscopic examinations and subjective and objective assessments. Epidemiological characteristics, accompanying organic and psychological disorders, preferred methods of treatment, and previous treatment outcomes were examined for each patient. A comparison was made based on voice disorders and responses to treatment between patients who received psychotherapy and patients who did not. Participants in this study comprised 58 patients, 10 male and 48 female. Voice therapy was applied in all patients, 54 (93.1%) of whom had improvement in their voice. Although all patients were advised to undergo psychological assessment, only 60.3% (35/58) of them underwent psychological assessment. No statistically significant difference was found between patients who did receive psychological support concerning their treatment responses and patients who did not. Relapse occurred in 14.7% (5/34) of the patients who applied for psychological assessment and in 50% (10/20) of those who did not. There was a statistically significant difference in relapse rates, which was higher among patients who did not receive psychological support (P < 0.005). Voice therapy is an efficient treatment method for PVD. However, in the long-term follow-up, relapse of the disease is observed to be higher among patients who failed to follow up on the recommendation for psychological assessment. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  2. Design and realization of intelligent tourism service system based on voice interaction

    NASA Astrophysics Data System (ADS)

    Hu, Lei-di; Long, Yi; Qian, Cheng-yang; Zhang, Ling; Lv, Guo-nian

    2008-10-01

    Voice technology is one of the important contents to improve the intelligence and humanization of tourism service system. Combining voice technology, the paper concentrates on application needs and the composition of system to present an overall intelligent tourism service system's framework consisting of presentation layer, Web services layer, and tourism application service layer. On the basis, the paper further elaborated the implementation of the system and its key technologies, including intelligent voice interactive technology, seamless integration technology of multiple data sources, location-perception-based guides' services technology, and tourism safety control technology. Finally, according to the situation of Nanjing tourism, a prototype of Tourism Services System is realized.

  3. A long distance voice transmission system based on the white light LED

    NASA Astrophysics Data System (ADS)

    Tian, Chunyu; Wei, Chang; Wang, Yulian; Wang, Dachi; Yu, Benli; Xu, Feng

    2017-10-01

    A long distance voice transmission system based on a visible light communication technology (VLCT) is proposed in the paper. Our proposed system includes transmitter, receiver and the voice signal processing of single chip microcomputer. In the compact-sized LED transmitter, we use on-off-keying and not-return-to-zero (OOK-NRZ) to easily realize high speed modulation, and then systematic complexity is reduced. A voice transmission system, which possesses the properties of the low-noise and wide modulation band, is achieved by the design of high efficiency receiving optical path and using filters to reduce noise from the surrounding light. To improve the speed of the signal processing, we use single chip microcomputer to code and decode voice signal. Furthermore, serial peripheral interface (SPI) is adopted to accurately transmit voice signal data. The test results of our proposed system show that the transmission distance of this system is more than100 meters with the maximum data rate of 1.5 Mbit/s and a SNR of 30dB. This system has many advantages, such as simple construction, low cost and strong practicality. Therefore, it has extensive application prospect in the fields of the emergency communication and indoor wireless communication, etc.

  4. Vocal and neural responses to unexpected changes in voice pitch auditory feedback during register transitions

    PubMed Central

    Patel, Sona; Lodhavia, Anjli; Frankford, Saul; Korzyukov, Oleg; Larson, Charles R.

    2016-01-01

    Objective/Hypothesis It is known that singers are able to control their voice to maintain a relatively constant vocal quality while transitioning between vocal registers; however, the neural mechanisms underlying this effect are not understood. It was hypothesized that greater attention to the acoustical feedback of the voice and increased control of the vocal musculature during register transitions compared to singing within a register would be represented as neurological differences in event-related potentials (ERPs). Study Design/Methods Nine singers sang musical notes at the high end of the modal register (the boundary between the modal and head/falsetto registers) and at the low end (the boundary between the modal and fry/pulse registers). While singing, the pitch of the voice auditory feedback was unexpectedly shifted either into the adjacent register (“toward” the register boundary) or within the modal register (“away from” the boundary). Singers were instructed to maintain a constant pitch and ignore any changes to their voice feedback. Results Vocal response latencies and magnitude of the accompanying N1 and P2 ERPs were greatest at the lower (modal-fry) boundary when the pitch shift carried the subjects’ voices into the fry register as opposed to remaining within the modal register. Conclusions These findings suggest that when a singer lowers the pitch of their voice such that it enters the fry register from the modal register, there is increased sensory-motor control of the voice, reflected as increased magnitude of the neural potentials to help minimize qualitative changes in the voice. PMID:26739860

  5. The Temporal Lobes Differentiate between the Voices of Famous and Unknown People: An Event-Related fMRI Study on Speaker Recognition

    PubMed Central

    Bethmann, Anja; Scheich, Henning; Brechmann, André

    2012-01-01

    It is widely accepted that the perception of human voices is supported by neural structures located along the superior temporal sulci. However, there is an ongoing discussion to what extent the activations found in fMRI studies are evoked by the vocal features themselves or are the result of phonetic processing. To show that the temporal lobes are indeed engaged in voice processing, short utterances spoken by famous and unknown people were presented to healthy young participants whose task it was to identify the familiar speakers. In two event-related fMRI experiments, the temporal lobes were found to differentiate between familiar and unfamiliar voices such that named voices elicited higher BOLD signal intensities than unfamiliar voices. Yet, the temporal cortices did not only discriminate between familiar and unfamiliar voices. Experiment 2, which required overtly spoken responses and allowed to distinguish between four familiarity grades, revealed that there was a fine-grained differentiation between all of these familiarity levels with higher familiarity being associated with larger BOLD signal amplitudes. Finally, we observed a gradual response change such that the BOLD signal differences between unfamiliar and highly familiar voices increased with the distance of an area from the transverse temporal gyri, especially towards the anterior temporal cortex and the middle temporal gyri. Therefore, the results suggest that (the anterior and non-superior portions of) the temporal lobes participate in voice-specific processing independent from phonetic components also involved in spoken speech material. PMID:23112826

  6. Speaker's comfort in teaching environments: voice problems in Swedish teaching staff.

    PubMed

    Åhlander, Viveka Lyberg; Rydell, Roland; Löfqvist, Anders

    2011-07-01

    The primary objective of this study was to examine how a group of Swedish teachers rate aspects of their working environment that can be presumed to have an impact on vocal behavior and voice problems. The secondary objective was to explore the prevalence of voice problems in Swedish teachers. Questionnaires were distributed to the teachers of 23 randomized schools. Teaching staff at all levels were included, except preschool teachers and teachers at specialized, vocational high schools. The response rate was 73%. The results showed that 13% of the whole group reported voice problems occurring sometimes, often, or always. The teachers reporting voice problems were compared with those without problems. There were significant differences among the groups for several items. The teachers with voice problems rated items on room acoustics and work environment as more noticeable. This group also reported voice symptoms, such as hoarseness, throat clearing, and voice change, to a significantly higher degree, even though teachers in both groups reported some voice symptoms. Absence from work because of voice problems was also significantly more common in the group with voice problems--35% versus 9% in the group without problems. We may conclude that teachers suffering from voice problems react stronger to loading factors in the teaching environment, report more frequent symptoms of voice discomfort, and are more often absent from work because of voice problems than their voice-healthy colleagues. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  7. Effects of the Voice over Internet Protocol on Perturbation Analysis of Normal and Pathological Phonation

    PubMed Central

    Zhu, Yanmei; Witt, Rachel E.; MacCallum, Julia K.; Jiang, Jack J.

    2010-01-01

    Objective In this study, a Voice over Internet Protocol (VoIP) communication based on G.729 protocol was simulated to determine the effects of this system on acoustic perturbation parameters of normal and pathological voice signals. Patients and Methods: Fifty recordings of normal voice and 48 recordings of pathological voice affected by laryngeal paralysis were transmitted through a VoIP communication system. The acoustic analysis programs of CSpeech and MDVP were used to determine the percent jitter and percent shimmer from the voice samples before and after VoIP transmission. The effects of three frequently used audio compression protocols (MP3, WMA, and FLAC) on the perturbation measures were also studied. Results It was found that VoIP transmission disrupts the waveform and increases the percent jitter and percent shimmer of voice samples. However, after VoIP transmission, significant discrimination between normal and pathological voices affected by laryngeal paralysis was still possible. It was found that the lossless compression method FLAC does not exert any influence on the perturbation measures. The lossy compression methods MP3 and WMA increase percent jitter and percent shimmer values. Conclusion This study validates the feasibility of these transmission and compression protocols in developing remote voice signal data collection and assessment systems. PMID:20588051

  8. Application of Mathematical Signal Processing Techniques to Mission Systems. (l’Application des techniques mathematiques du traitement du signal aux systemes de conduite des missions)

    DTIC Science & Technology

    1999-11-01

    represents the linear time invariant (LTI) response of the combined analysis /synthesis system while the second repre- sents the aliasing introduced into...effectively to implement voice scrambling systems based on time - frequency permutation . The most general form of such a system is shown in Fig. 22 where...92201 NEUILLY-SUR-SEINE CEDEX, FRANCE RTO LECTURE SERIES 216 Application of Mathematical Signal Processing Techniques to Mission Systems (1

  9. Apollo experience report: Voice communications techniques and performance

    NASA Technical Reports Server (NTRS)

    Dabbs, J. H.; Schmidt, O. L.

    1972-01-01

    The primary performance requirement of the spaceborne Apollo voice communications system is percent word intelligibility, which is related to other link/channel parameters. The effect of percent word intelligibility on voice channel design and a description of the verification procedures are included. Development and testing performance problems and the techniques used to solve the problems are also discussed. Voice communications performance requirements should be comprehensive and verified easily; the total system must be considered in component design, and the necessity of voice processing and the associated effect on noise, distortion, and cross talk should be examined carefully.

  10. Vocal Fold Bowing in Elderly Male Monozygotic Twins: A Case Study

    PubMed Central

    Tanner, Kristine; Sauder, Cara; Thibeault, Susan L.; Dromey, Christopher; Smith, Marshall E.

    2009-01-01

    Objectives This study examined case histories, diagnostic features, and treatment response in two 79-year-old male monozygotic (identical) twins with vocal fold bowing, exploring both genetic and environmental factors. Study Design Case study. Methods DNA concordance was examined via cheek swab. Case histories, videostroboscopy, auditory- and visual-perceptual assessment, electromyography, acoustic measures, and Voice Handicap ratings were undertaken. Both twins underwent surgical intervention and subsequent voice therapy. Results Monozygosity was confirmed for DNA polymorphisms, with 10 of 10 concordance for STR DNA markers. For both twins, auditory and visual-perceptual assessments indicated severe bowing, hoarseness and breathiness, although Twin 1 was judged to be extremely severe. Differences in RMS amplitudes were observed for TA and LCA muscles, with smaller relative amplitudes observed for the Twin 1 versus Twin 2. No consistent voice improvement was observed following surgical intervention(s), despite improved mid-membranous vocal fold closure. Marked reductions in Voice Handicap Index total scores were observed following behavioral voice therapy, coinciding with increased mid-membranous and posterior laryngeal (interarytenoid) glottal closure. No substantive differences in acoustic measures were observed. Conclusions Vocal fold bowing was more severe for Twin 1 versus Twin 2 despite identical heritability factors. Overall voice improvement with treatment was greater for Twin 2 than Twin 1. Environmental factors might partially account for the differences observed between the twins, including variability in their responsiveness to behavioral voice therapy. Voice therapy was useful in improving mid-membranous and posterior laryngeal closure, although dysphonia remained severe in both cases. PMID:19664899

  11. Opposing and following responses in sensorimotor speech control: Why responses go both ways.

    PubMed

    Franken, Matthias K; Acheson, Daniel J; McQueen, James M; Hagoort, Peter; Eisner, Frank

    2018-06-04

    When talking, speakers continuously monitor and use the auditory feedback of their own voice to control and inform speech production processes. When speakers are provided with auditory feedback that is perturbed in real time, most of them compensate for this by opposing the feedback perturbation. But some responses follow the perturbation. In the present study, we investigated whether the state of the speech production system at perturbation onset may determine what type of response (opposing or following) is made. The results suggest that whether a perturbation-related response is opposing or following depends on ongoing fluctuations of the production system: The system initially responds by doing the opposite of what it was doing. This effect and the nontrivial proportion of following responses suggest that current production models are inadequate: They need to account for why responses to unexpected sensory feedback depend on the production system's state at the time of perturbation.

  12. Towards an Adolescent Friendly Methodology: Accessing the Authentic through Collective Reflection

    ERIC Educational Resources Information Center

    Keeffe, Mary; Andrews, Dorothy

    2015-01-01

    The re-emergence of student voice presents a challenge to schools and researchers to become more responsive to the voice of adolescents in education and in research. However, the poor articulation of the nature of student voice to date is confirmation of the complex and important nature of the personal advocacy and human agency that is involved in…

  13. Examining Response to a One-to-One Computer Initiative: Student and Teacher Voices

    ERIC Educational Resources Information Center

    Storz, Mark G.; Hoffman, Amy R.

    2013-01-01

    The impact of a one-to-one computing initiative at a Midwestern urban middle school was examined through phenomenological research techniques focusing on the voices of eighth grade students and their teachers. Analysis of transcripts from pre and post-implementation interviews of 47 students and eight teachers yielded patterns of responses to…

  14. They Are Talking: Are We Listening? Using Student Voice to Enhance Culturally Responsive Teaching

    ERIC Educational Resources Information Center

    Anderson, Gina; Cowart, Melinda

    2012-01-01

    This conversational report uses student voice as data to determine whether the culture of urban sixth graders is being acknowledged and valued in the curriculum. While culturally responsive teaching has been touted by scholars as an important aspect of multicultural education and curriculum reform for at least a decade, students have seldom been…

  15. Multi-modal assessment of on-road demand of voice and manual phone calling and voice navigation entry across two embedded vehicle systems.

    PubMed

    Mehler, Bruce; Kidd, David; Reimer, Bryan; Reagan, Ian; Dobres, Jonathan; McCartt, Anne

    2016-03-01

    One purpose of integrating voice interfaces into embedded vehicle systems is to reduce drivers' visual and manual distractions with 'infotainment' technologies. However, there is scant research on actual benefits in production vehicles or how different interface designs affect attentional demands. Driving performance, visual engagement, and indices of workload (heart rate, skin conductance, subjective ratings) were assessed in 80 drivers randomly assigned to drive a 2013 Chevrolet Equinox or Volvo XC60. The Chevrolet MyLink system allowed completing tasks with one voice command, while the Volvo Sensus required multiple commands to navigate the menu structure. When calling a phone contact, both voice systems reduced visual demand relative to the visual-manual interfaces, with reductions for drivers in the Equinox being greater. The Equinox 'one-shot' voice command showed advantages during contact calling but had significantly higher error rates than Sensus during destination address entry. For both secondary tasks, neither voice interface entirely eliminated visual demand. Practitioner Summary: The findings reinforce the observation that most, if not all, automotive auditory-vocal interfaces are multi-modal interfaces in which the full range of potential demands (auditory, vocal, visual, manipulative, cognitive, tactile, etc.) need to be considered in developing optimal implementations and evaluating drivers' interaction with the systems. Social Media: In-vehicle voice-interfaces can reduce visual demand but do not eliminate it and all types of demand need to be taken into account in a comprehensive evaluation.

  16. Multi-modal assessment of on-road demand of voice and manual phone calling and voice navigation entry across two embedded vehicle systems

    PubMed Central

    Mehler, Bruce; Kidd, David; Reimer, Bryan; Reagan, Ian; Dobres, Jonathan; McCartt, Anne

    2016-01-01

    Abstract One purpose of integrating voice interfaces into embedded vehicle systems is to reduce drivers’ visual and manual distractions with ‘infotainment’ technologies. However, there is scant research on actual benefits in production vehicles or how different interface designs affect attentional demands. Driving performance, visual engagement, and indices of workload (heart rate, skin conductance, subjective ratings) were assessed in 80 drivers randomly assigned to drive a 2013 Chevrolet Equinox or Volvo XC60. The Chevrolet MyLink system allowed completing tasks with one voice command, while the Volvo Sensus required multiple commands to navigate the menu structure. When calling a phone contact, both voice systems reduced visual demand relative to the visual–manual interfaces, with reductions for drivers in the Equinox being greater. The Equinox ‘one-shot’ voice command showed advantages during contact calling but had significantly higher error rates than Sensus during destination address entry. For both secondary tasks, neither voice interface entirely eliminated visual demand. Practitioner Summary: The findings reinforce the observation that most, if not all, automotive auditory–vocal interfaces are multi-modal interfaces in which the full range of potential demands (auditory, vocal, visual, manipulative, cognitive, tactile, etc.) need to be considered in developing optimal implementations and evaluating drivers’ interaction with the systems. Social Media: In-vehicle voice-interfaces can reduce visual demand but do not eliminate it and all types of demand need to be taken into account in a comprehensive evaluation. PMID:26269281

  17. White House Communications Agency (WHCA) Presidential Voice Communications Rack Mount System Mechanical Drawing Package

    DTIC Science & Technology

    2015-12-01

    Rack Mount System Mechanical Drawing Package by Steven P Callaway Approved for public release; distribution unlimited...Laboratory White House Communications Agency (WHCA) Presidential Voice Communications Rack Mount System Mechanical Drawing Package by Steven P...Note 3. DATES COVERED (From - To) 04/2013 4. TITLE AND SUBTITLE White House Communications Agency (WHCA) Presidential Voice Communications Rack

  18. Central Nervous System Control of Voice and Swallowing

    PubMed Central

    Ludlow, Christy L.

    2015-01-01

    This review of the central nervous control systems for voice and swallowing has suggested that the traditional concepts of a separation between cortical and limbic and brain stem control should be refined and more integrative. For voice production, a separation of the non-human vocalization system from the human learned voice production system has been posited based primarily on studies of non-human primates. However, recent humans studies of emotionally based vocalizations and human volitional voice production has shown more integration between these two systems than previously proposed. Recent human studies have shown that reflexive vocalization as well as learned voice production not involving speech, involve a common integrative system. On the other hand, recent studies of non-human primates have provided evidence of some cortical activity during vocalization and cortical changes with training during vocal behavior. For swallowing, evidence from the macaque and functional brain imaging in humans indicates that the control for the pharyngeal phase of swallowing is not primarily under brain stem mechanisms as previously proposed. Studies suggest that the initiation and patterning of swallowing for the pharyngeal phase is also under active cortical control for both spontaneous as well as volitional swallowing in awake humans and non-human primates. PMID:26241238

  19. Neurobiological correlates of emotional intelligence in voice and face perception networks

    PubMed Central

    Karle, Kathrin N; Ethofer, Thomas; Jacob, Heike; Brück, Carolin; Erb, Michael; Lotze, Martin; Nizielski, Sophia; Schütz, Astrid; Wildgruber, Dirk; Kreifelts, Benjamin

    2018-01-01

    Abstract Facial expressions and voice modulations are among the most important communicational signals to convey emotional information. The ability to correctly interpret this information is highly relevant for successful social interaction and represents an integral component of emotional competencies that have been conceptualized under the term emotional intelligence. Here, we investigated the relationship of emotional intelligence as measured with the Salovey-Caruso-Emotional-Intelligence-Test (MSCEIT) with cerebral voice and face processing using functional and structural magnetic resonance imaging. MSCEIT scores were positively correlated with increased voice-sensitivity and gray matter volume of the insula accompanied by voice-sensitivity enhanced connectivity between the insula and the temporal voice area, indicating generally increased salience of voices. Conversely, in the face processing system, higher MSCEIT scores were associated with decreased face-sensitivity and gray matter volume of the fusiform face area. Taken together, these findings point to an alteration in the balance of cerebral voice and face processing systems in the form of an attenuated face-vs-voice bias as one potential factor underpinning emotional intelligence. PMID:29365199

  20. Neurobiological correlates of emotional intelligence in voice and face perception networks.

    PubMed

    Karle, Kathrin N; Ethofer, Thomas; Jacob, Heike; Brück, Carolin; Erb, Michael; Lotze, Martin; Nizielski, Sophia; Schütz, Astrid; Wildgruber, Dirk; Kreifelts, Benjamin

    2018-02-01

    Facial expressions and voice modulations are among the most important communicational signals to convey emotional information. The ability to correctly interpret this information is highly relevant for successful social interaction and represents an integral component of emotional competencies that have been conceptualized under the term emotional intelligence. Here, we investigated the relationship of emotional intelligence as measured with the Salovey-Caruso-Emotional-Intelligence-Test (MSCEIT) with cerebral voice and face processing using functional and structural magnetic resonance imaging. MSCEIT scores were positively correlated with increased voice-sensitivity and gray matter volume of the insula accompanied by voice-sensitivity enhanced connectivity between the insula and the temporal voice area, indicating generally increased salience of voices. Conversely, in the face processing system, higher MSCEIT scores were associated with decreased face-sensitivity and gray matter volume of the fusiform face area. Taken together, these findings point to an alteration in the balance of cerebral voice and face processing systems in the form of an attenuated face-vs-voice bias as one potential factor underpinning emotional intelligence.

  1. Literature review of voice recognition and generation technology for Army helicopter applications

    NASA Astrophysics Data System (ADS)

    Christ, K. A.

    1984-08-01

    This report is a literature review on the topics of voice recognition and generation. Areas covered are: manual versus vocal data input, vocabulary, stress and workload, noise, protective masks, feedback, and voice warning systems. Results of the studies presented in this report indicate that voice data entry has less of an impact on a pilot's flight performance, during low-level flying and other difficult missions, than manual data entry. However, the stress resulting from such missions may cause the pilot's voice to change, reducing the recognition accuracy of the system. The noise present in helicopter cockpits also causes the recognition accuracy to decrease. Noise-cancelling devices are being developed and improved upon to increase the recognition performance in noisy environments. Future research in the fields of voice recognition and generation should be conducted in the areas of stress and workload, vocabulary, and the types of voice generation best suited for the helicopter cockpit. Also, specific tasks should be studied to determine whether voice recognition and generation can be effectively applied.

  2. Contemplating what I would do if someone got in my house: intentions of older homebound women living alone.

    PubMed

    Porter, Eileen J

    2008-01-01

    There is little research guiding interventions to help old homebound women prepare to manage an intrusion event. During a phenomenological study of the experience of reaching help quickly, I compared intentions during a possible intrusion event for 9 women subscribing to a personal emergency response system and 5 nonsubscribers. The phenomenon of contemplating what I would do if an intruder got in my home had 4 components. Only 2 personal emergency response system subscribers voiced the definitive intention to use the personal emergency response system. Findings underpin a new empirical perspective of competence grounded in situations relevant to living alone at home rather than specific tasks of daily living.

  3. Voice rest after vocal fold surgery: current practice and evidence.

    PubMed

    Coombs, A C; Carswell, A J; Tierney, P A

    2013-08-01

    Voice rest is commonly recommended after vocal fold surgery, but there is a lack of evidence base and no standard protocol. The aim of this study was to establish common practice regarding voice rest following vocal fold surgery. An online survey was circulated via e-mail invitation to members of the ENT UK Expert Panel between October and November 2011. The survey revealed that 86.5 per cent of respondents agreed that 'complete voice rest' means no sound production at all, but there was variability in how 'relative voice rest' was defined. There was no dominant type of voice rest routinely recommended after surgery for laryngeal papillomatosis or intermediate pathologies. There was considerable variability in the duration of voice rest recommended, with no statistically significant, most popular response (except for malignant lesions). Surgeons with less than 10 years of experience were more likely to recommend fewer days of voice rest. There is a lack of consistency in advice given to patients after vocal fold surgery, in terms of both type and length of voice rest. This may arise from an absence of robust evidence on which to base practice.

  4. Overgeneral autobiographical memory bias in clinical and non-clinical voice hearers.

    PubMed

    Jacobsen, Pamela; Peters, Emmanuelle; Ward, Thomas; Garety, Philippa A; Jackson, Mike; Chadwick, Paul

    2018-03-14

    Hearing voices can be a distressing and disabling experience for some, whilst it is a valued experience for others, so-called 'healthy voice-hearers'. Cognitive models of psychosis highlight the role of memory, appraisal and cognitive biases in determining emotional and behavioural responses to voices. A memory bias potentially associated with distressing voices is the overgeneral memory bias (OGM), namely the tendency to recall a summary of events rather than specific occasions. It may limit access to autobiographical information that could be helpful in re-appraising distressing experiences, including voices. We investigated the possible links between OGM and distressing voices in psychosis by comparing three groups: (1) clinical voice-hearers (N = 39), (2) non-clinical voice-hearers (N = 35) and (3) controls without voices (N = 77) on a standard version of the autobiographical memory test (AMT). Clinical and non-clinical voice-hearers also completed a newly adapted version of the task, designed to assess voices-related memories (vAMT). As hypothesised, the clinical group displayed an OGM bias by retrieving fewer specific autobiographical memories on the AMT compared with both the non-clinical and control groups, who did not differ from each other. The clinical group also showed an OGM bias in recall of voice-related memories on the vAMT, compared with the non-clinical group. Clinical voice-hearers display an OGM bias when compared with non-clinical voice-hearers on both general and voices-specific recall tasks. These findings have implications for the refinement and targeting of psychological interventions for psychosis.

  5. Uncertainty quantification of voice signal production mechanical model and experimental updating

    NASA Astrophysics Data System (ADS)

    Cataldo, E.; Soize, C.; Sampaio, R.

    2013-11-01

    The aim of this paper is to analyze the uncertainty quantification in a voice production mechanical model and update the probability density function corresponding to the tension parameter using the Bayes method and experimental data. Three parameters are considered uncertain in the voice production mechanical model used: the tension parameter, the neutral glottal area and the subglottal pressure. The tension parameter of the vocal folds is mainly responsible for the changing of the fundamental frequency of a voice signal, generated by a mechanical/mathematical model for producing voiced sounds. The three uncertain parameters are modeled by random variables. The probability density function related to the tension parameter is considered uniform and the probability density functions related to the neutral glottal area and the subglottal pressure are constructed using the Maximum Entropy Principle. The output of the stochastic computational model is the random voice signal and the Monte Carlo method is used to solve the stochastic equations allowing realizations of the random voice signals to be generated. For each realization of the random voice signal, the corresponding realization of the random fundamental frequency is calculated and the prior pdf of this random fundamental frequency is then estimated. Experimental data are available for the fundamental frequency and the posterior probability density function of the random tension parameter is then estimated using the Bayes method. In addition, an application is performed considering a case with a pathology in the vocal folds. The strategy developed here is important mainly due to two things. The first one is related to the possibility of updating the probability density function of a parameter, the tension parameter of the vocal folds, which cannot be measured direct and the second one is related to the construction of the likelihood function. In general, it is predefined using the known pdf. Here, it is constructed in a new and different manner, using the own system considered.

  6. Talking Wheelchair

    NASA Technical Reports Server (NTRS)

    1981-01-01

    Communication is made possible for disabled individuals by means of an electronic system, developed at Stanford University's School of Medicine, which produces highly intelligible synthesized speech. Familiarly known as the "talking wheelchair" and formally as the Versatile Portable Speech Prosthesis (VPSP). Wheelchair mounted system consists of a word processor, a video screen, a voice synthesizer and a computer program which instructs the synthesizer how to produce intelligible sounds in response to user commands. Computer's memory contains 925 words plus a number of common phrases and questions. Memory can also store several thousand other words of the user's choice. Message units are selected by operating a simple switch, joystick or keyboard. Completed message appears on the video screen, then user activates speech synthesizer, which generates a voice with a somewhat mechanical tone. With the keyboard, an experienced user can construct messages as rapidly as 30 words per minute.

  7. Impact of a voice recognition system on report cycle time and radiologist reading time

    NASA Astrophysics Data System (ADS)

    Melson, David L.; Brophy, Robert; Blaine, G. James; Jost, R. Gilbert; Brink, Gary S.

    1998-07-01

    Because of its exciting potential to improve clinical service, as well as reduce costs, a voice recognition system for radiological dictation was recently installed at our institution. This system will be clinically successful if it dramatically reduces radiology report turnaround time without substantially affecting radiologist dictation and editing time. This report summarizes an observer study currently under way in which radiologist reporting times using the traditional transcription system and the voice recognition system are compared. Four radiologists are observed interpreting portable intensive care unit (ICU) chest examinations at a workstation in the chest reading area. Data are recorded with the radiologists using the transcription system and using the voice recognition system. The measurements distinguish between time spent performing clerical tasks and time spent actually dictating the report. Editing time and the number of corrections made are recorded. Additionally, statistics are gathered to assess the voice recognition system's impact on the report cycle time -- the time from report dictation to availability of an edited and finalized report -- and the length of reports.

  8. Analysis of the Auditory Feedback and Phonation in Normal Voices.

    PubMed

    Arbeiter, Mareike; Petermann, Simon; Hoppe, Ulrich; Bohr, Christopher; Doellinger, Michael; Ziethe, Anke

    2018-02-01

    The aim of this study was to investigate the auditory feedback mechanisms and voice quality during phonation in response to a spontaneous pitch change in the auditory feedback. Does the pitch shift reflex (PSR) change voice pitch and voice quality? Quantitative and qualitative voice characteristics were analyzed during the PSR. Twenty-eight healthy subjects underwent transnasal high-speed video endoscopy (HSV) at 8000 fps during sustained phonation [a]. While phonating, the subjects heard their sound pitched up for 700 cents (interval of a fifth), lasting 300 milliseconds in their auditory feedback. The electroencephalography (EEG), acoustic voice signal, electroglottography (EGG), and high-speed-videoendoscopy (HSV) were analyzed to compare feedback mechanisms for the pitched and unpitched condition of the phonation paradigm statistically. Furthermore, quantitative and qualitative voice characteristics were analyzed. The PSR was successfully detected within all signals of the experimental tools (EEG, EGG, acoustic voice signal, HSV). A significant increase of the perturbation measures and an increase of the values of the acoustic parameters during the PSR were observed, especially for the audio signal. The auditory feedback mechanism seems not only to control for voice pitch but also for voice quality aspects.

  9. Accelerometer-based automatic voice onset detection in speech mapping with navigated repetitive transcranial magnetic stimulation.

    PubMed

    Vitikainen, Anne-Mari; Mäkelä, Elina; Lioumis, Pantelis; Jousmäki, Veikko; Mäkelä, Jyrki P

    2015-09-30

    The use of navigated repetitive transcranial magnetic stimulation (rTMS) in mapping of speech-related brain areas has recently shown to be useful in preoperative workflow of epilepsy and tumor patients. However, substantial inter- and intraobserver variability and non-optimal replicability of the rTMS results have been reported, and a need for additional development of the methodology is recognized. In TMS motor cortex mappings the evoked responses can be quantitatively monitored by electromyographic recordings; however, no such easily available setup exists for speech mappings. We present an accelerometer-based setup for detection of vocalization-related larynx vibrations combined with an automatic routine for voice onset detection for rTMS speech mapping applying naming. The results produced by the automatic routine were compared with the manually reviewed video-recordings. The new method was applied in the routine navigated rTMS speech mapping for 12 consecutive patients during preoperative workup for epilepsy or tumor surgery. The automatic routine correctly detected 96% of the voice onsets, resulting in 96% sensitivity and 71% specificity. Majority (63%) of the misdetections were related to visible throat movements, extra voices before the response, or delayed naming of the previous stimuli. The no-response errors were correctly detected in 88% of events. The proposed setup for automatic detection of voice onsets provides quantitative additional data for analysis of the rTMS-induced speech response modifications. The objectively defined speech response latencies increase the repeatability, reliability and stratification of the rTMS results. Copyright © 2015 Elsevier B.V. All rights reserved.

  10. Mobile phone-based interactive voice response as a tool for improving access to healthcare in remote areas in Ghana - an evaluation of user experiences.

    PubMed

    Brinkel, J; May, J; Krumkamp, R; Lamshöft, M; Kreuels, B; Owusu-Dabo, E; Mohammed, A; Bonacic Marinovic, A; Dako-Gyeke, P; Krämer, A; Fobil, J N

    2017-05-01

    To investigate and determine the factors that enhanced or constituted barriers to the acceptance of an mHealth system which was piloted in Asante-Akim North District of Ghana to support healthcare of children. Four semi-structured focus group discussions were conducted with a total of 37 mothers. Participants were selected from a study population of mothers who subscribed to a pilot mHealth system which used an interactive voice response (IVR) for its operations. Data were evaluated using qualitative content analysis methods. In addition, a short quantitative questionnaire assessed system's usability (SUS). Results revealed 10 categories of factors that facilitated user acceptance of the IVR system including quality-of-care experience, health education and empowerment of women. The eight categories of factors identified as barriers to user acceptance included the lack of human interaction, lack of update and training on the electronic advices provided and lack of social integration of the system into the community. The usability (SUS median: 79.3; range: 65-97.5) of the system was rated acceptable. The principles of the tested mHealth system could be of interest during infectious disease outbreaks, such as Ebola or Lassa fever, when there might be a special need for disease-specific health information within populations. © 2017 John Wiley & Sons Ltd.

  11. Vibrant Student Voices: Exploring Effects of the Use of Clickers in Large College Courses

    ERIC Educational Resources Information Center

    Hoekstra, Angel

    2008-01-01

    Teachers have begun using student response systems (SRSs) in an effort to enhance the learning process in higher education courses. Research providing detailed information about how interactive technologies affect students as they learn is crucial for professors who seek to improve teaching quality, attendance rates and student learning. This…

  12. Understanding and Developing Interactive Voice Response Systems to Support Online Engagement of Older Adults

    ERIC Educational Resources Information Center

    Brewer, Robin Nicole

    2017-01-01

    Increasingly, people are engaging online and can participate in activities like searching for information, communicating with family and friends, and self-expression. However, some populations such as older adults, face barriers to online participation like device cost, access, and learnability, which prevent them from reaping the benefits of…

  13. Using interactive voice response to improve disease management and compliance with acute coronary syndrome best practice guidelines: A randomized controlled trial.

    PubMed

    Sherrard, Heather; Duchesne, Lloyd; Wells, George; Kearns, Sharon Ann; Struthers, Christine

    2015-01-01

    There is evidence from large clinical trials that compliance with standardized best practice guidelines (BPGs) improves survival of acute coronary syndrome (ACS) patients. However, their application is often suboptimal. In this study, the researchers evaluated whether the use of an interactive voice response (IVR) follow-up system improved ACS BPG compliance. This was a single-centre randomized control trial (RCT) of 1,608 patients (IVR=803; usual care=805). The IVR group received five automated calls in 12 months. The primary composite outcome was increased medication compliance and decreased adverse events. A significant improvement of 60% in the IVR group for the primary composite outcome was found (RR 1.60, 95% CI: 1.29 to 2.00, p <0.001). There was significant improvement in medication compliance (p <0.001) and decrease in unplanned medical visits (p = 0.023). At one year, the majority of patients ( 85%) responded positively to using the system again. Follow-up by IVR produced positive outcomes in ACS patients.

  14. Voices to reckon with: perceptions of voice identity in clinical and non-clinical voice hearers

    PubMed Central

    Badcock, Johanna C.; Chhabra, Saruchi

    2013-01-01

    The current review focuses on the perception of voice identity in clinical and non-clinical voice hearers. Identity perception in auditory verbal hallucinations (AVH) is grounded in the mechanisms of human (i.e., real, external) voice perception, and shapes the emotional (distress) and behavioral (help-seeking) response to the experience. Yet, the phenomenological assessment of voice identity is often limited, for example to the gender of the voice, and has failed to take advantage of recent models and evidence on human voice perception. In this paper we aim to synthesize the literature on identity in real and hallucinated voices and begin by providing a comprehensive overview of the features used to judge voice identity in healthy individuals and in people with schizophrenia. The findings suggest some subtle, but possibly systematic biases across different levels of voice identity in clinical hallucinators that are associated with higher levels of distress. Next we provide a critical evaluation of voice processing abilities in clinical and non-clinical voice hearers, including recent data collected in our laboratory. Our studies used diverse methods, assessing recognition and binding of words and voices in memory as well as multidimensional scaling of voice dissimilarity judgments. The findings overall point to significant difficulties recognizing familiar speakers and discriminating between unfamiliar speakers in people with schizophrenia, both with and without AVH. In contrast, these voice processing abilities appear to be generally intact in non-clinical hallucinators. The review highlights some important avenues for future research and treatment of AVH associated with a need for care, and suggests some novel insights into other symptoms of psychosis. PMID:23565088

  15. 14 CFR 25.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 14 Aeronautics and Space 1 2014-01-01 2014-01-01 false Cockpit voice recorders. 25.1457 Section 25... recorders. (a) Each cockpit voice recorder required by the operating rules of this chapter must be approved... interphone system. (4) Voice or audio signals identifying navigation or approach aids introduced into a...

  16. 14 CFR 25.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 14 Aeronautics and Space 1 2013-01-01 2013-01-01 false Cockpit voice recorders. 25.1457 Section 25... recorders. (a) Each cockpit voice recorder required by the operating rules of this chapter must be approved... interphone system. (4) Voice or audio signals identifying navigation or approach aids introduced into a...

  17. 14 CFR 29.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 14 Aeronautics and Space 1 2012-01-01 2012-01-01 false Cockpit voice recorders. 29.1457 Section 29... recorders. (a) Each cockpit voice recorder required by the operating rules of this chapter must be approved... interphone system. (4) Voice or audio signals identifying navigation or approach aids introduced into a...

  18. 14 CFR 29.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 14 Aeronautics and Space 1 2013-01-01 2013-01-01 false Cockpit voice recorders. 29.1457 Section 29... recorders. (a) Each cockpit voice recorder required by the operating rules of this chapter must be approved... interphone system. (4) Voice or audio signals identifying navigation or approach aids introduced into a...

  19. 14 CFR 25.1457 - Cockpit voice recorders.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 14 Aeronautics and Space 1 2012-01-01 2012-01-01 false Cockpit voice recorders. 25.1457 Section 25... recorders. (a) Each cockpit voice recorder required by the operating rules of this chapter must be approved... interphone system. (4) Voice or audio signals identifying navigation or approach aids introduced into a...

  20. Voice disorders in teachers: occupational risk factors and psycho-emotional factors.

    PubMed

    van Houtte, Evelyne; Claeys, Sofie; Wuyts, Floris; van Lierde, Kristiane

    2012-10-01

    Teaching is a high-risk occupation for developing voice disorders. The purpose of this study was to investigate previously described vocal risk factors as well as to identify new risk factors related to both the personal life of the teacher (fluid intake, voice-demanding activities, family history of voice disorders, and children at home) and to environmental factors (temperature changes, chalk use, presence of curtains, carpet, or air-conditioning, acoustics in the classroom, and noise in and outside the classroom). The study group comprised 994 teachers (response rate 46.6%). All participants completed a questionnaire. Chi-square tests and logistic regression analyses were performed. A total of 51.2% (509/994) of the teachers presented with voice disorders. Women reported more voice disorders compared to men (56.4% versus 40.4%, P < 0.001). Vocal risk factors were a family history of voice disorders (P = 0.005), temperature changes in the classroom (P = 0.017), the number of pupils per classroom (P = 0.001), and noise level inside the classroom (P = 0.001). Teachers with voice disorders presented a higher level of psychological distress (P < 0.001) compared to teachers without voice problems. Voice disorders are frequent among teachers, especially in female teachers. The results of this study emphasize that multiple factors are involved in the development of voice disorders.

  1. Automatic speech recognition in air-ground data link

    NASA Technical Reports Server (NTRS)

    Armstrong, Herbert B.

    1989-01-01

    In the present air traffic system, information presented to the transport aircraft cockpit crew may originate from a variety of sources and may be presented to the crew in visual or aural form, either through cockpit instrument displays or, most often, through voice communication. Voice radio communications are the most error prone method for air-ground data link. Voice messages can be misstated or misunderstood and radio frequency congestion can delay or obscure important messages. To prevent proliferation, a multiplexed data link display can be designed to present information from multiple data link sources on a shared cockpit display unit (CDU) or multi-function display (MFD) or some future combination of flight management and data link information. An aural data link which incorporates an automatic speech recognition (ASR) system for crew response offers several advantages over visual displays. The possibility of applying ASR to the air-ground data link was investigated. The first step was to review current efforts in ASR applications in the cockpit and in air traffic control and evaluated their possible data line application. Next, a series of preliminary research questions is to be developed for possible future collaboration.

  2. Motorcycle Start-stop System based on Intelligent Biometric Voice Recognition

    NASA Astrophysics Data System (ADS)

    Winda, A.; E Byan, W. R.; Sofyan; Armansyah; Zariantin, D. L.; Josep, B. G.

    2017-03-01

    Current mechanical key in the motorcycle is prone to bulgary, being stolen or misplaced. Intelligent biometric voice recognition as means to replace this mechanism is proposed as an alternative. The proposed system will decide whether the voice is belong to the user or not and the word utter by the user is ‘On’ or ‘Off’. The decision voice will be sent to Arduino in order to start or stop the engine. The recorded voice is processed in order to get some features which later be used as input to the proposed system. The Mel-Frequency Ceptral Coefficient (MFCC) is adopted as a feature extraction technique. The extracted feature is the used as input to the SVM-based identifier. Experimental results confirm the effectiveness of the proposed intelligent voice recognition and word recognition system. It show that the proposed method produces a good training and testing accuracy, 99.31% and 99.43%, respectively. Moreover, the proposed system shows the performance of false rejection rate (FRR) and false acceptance rate (FAR) accuracy of 0.18% and 17.58%, respectively. In the intelligent word recognition shows that the training and testing accuracy are 100% and 96.3%, respectively.

  3. Network Speech Systems Technology Program

    NASA Astrophysics Data System (ADS)

    Weinstein, C. J.

    1980-09-01

    This report documents work performed during FY 1980 on the DCA-sponsored Network Speech Systems Technology Program. The areas of work reported are: (1) communication systems studies in Demand-Assignment Multiple Access (DAMA), voice/data integration, and adaptive routing, in support of the evolving Defense Communications System (DCS) and Defense Switched Network (DSN); (2) a satellite/terrestrial integration design study including the functional design of voice and data interfaces to interconnect terrestrial and satellite network subsystems; and (3) voice-conferencing efforts dealing with support of the Secure Voice and Graphics Conferencing (SVGC) Test and Evaluation Program. Progress in definition and planning of experiments for the Experimental Integrated Switched Network (EISN) is detailed separately in an FY 80 Experiment Plan Supplement.

  4. Performer's attitudes toward seeking health care for voice issues: understanding the barriers.

    PubMed

    Gilman, Marina; Merati, Albert L; Klein, Adam M; Hapner, Edie R; Johns, Michael M

    2009-03-01

    Contemporary commercial music (CCM) performers rely heavily on their voice, yet may not be aware of the importance of proactive voice care. This investigation intends to identify perceptions and barriers to seeking voice care among CCM artists. This cross-sectional observational study used a 10-item Likert-based response questionnaire to assess current perceptions regarding voice care in a population of randomly selected participants of professional CCM conference. Subjects (n=78) were queried regarding their likelihood to seek medical care for minor medical problems and specifically problems with their voice. Additional questions investigated anxiety about seeking voice care from a physician specialist, speech language pathologist, or voice coach; apprehension regarding findings of laryngeal examination, laryngeal imaging procedures; and the effect of medical insurance on the likelihood of seeking medical care. Eighty-two percent of subjects reported that their voice was a critical part of their profession; 41% stated that they were not likely to seek medical care for problems with their voice; and only 19% were reluctant to seek care for general medical problems (P<0.001). Anxiety about seeking a clinician regarding their voice was not a deterrent. Most importantly, 39% of subjects do not seek medical attention for their voice problems due to medical insurance coverage. The CCM artists are less likely to seek medical care for voice problems compared with general medical problems. Availability of medical insurance may be a factor. Availability of affordable voice care and education about the importance of voice care is needed in this population of vocal performers.

  5. The Effect of Hydration on the Voice Quality of Future Professional Vocal Performers.

    PubMed

    van Wyk, Liezl; Cloete, Mariaan; Hattingh, Danel; van der Linde, Jeannie; Geertsema, Salome

    2017-01-01

    The application of systemic hydration as an instrument for optimal voice quality has been a common practice by several professional voice users over the years. Although the physiological action has been determined, the benefits on acoustic and perceptual characteristics are relatively unknown. The present study aimed to determine whether systemic hydration has beneficial outcomes on the voice quality of future professional voice users. A within-subject, pretest posttest design is applied to determine quantitative research results of female singing students between 18 and 32 years of age without a history of voice pathology. Acoustic and perceptual data were collected before and after a 2-hour singing rehearsal. The difference between the hypohydrated condition (controlled) and the hydrated condition (experimental) and the relationship between adequate hydration and acoustic and perceptual parameters of voice was then investigated. A statistical significant (P = 0.041) increase in jitter values were obtained for the hypohydrated condition. Increased maximum phonation time (MPT/z/) and higher maximum frequency for hydration indicated further statistical significant changes in voice quality (P = 0.028 and P = 0.015, respectively). Systemic hydration has positive outcomes on perceptual and acoustic parameters of voice quality for future professional singers. The singer's ability to sustain notes for longer and reach higher frequencies may reflect well in performances. Any positive change in voice quality may benefit the singer's occupational success and subsequently their social, emotional, and vocational well-being. More research evidence is needed to determine the parameters for implementing adequate hydration in vocal hygiene programs. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  6. A voice-actuated wind tunnel model leak checking system

    NASA Technical Reports Server (NTRS)

    Larson, William E.

    1989-01-01

    A computer program has been developed that improves the efficiency of wind tunnel model leak checking. The program uses a voice recognition unit to relay a technician's commands to the computer. The computer, after receiving a command, can respond to the technician via a voice response unit. Information about the model pressure orifice being checked is displayed on a gas-plasma terminal. On command, the program records up to 30 seconds of pressure data. After the recording is complete, the raw data and a straight line fit of the data are plotted on the terminal. This allows the technician to make a decision on the integrity of the orifice being checked. All results of the leak check program are stored in a database file that can be listed on the line printer for record keeping purposes or displayed on the terminal to help the technician find unchecked orifices. This program allows one technician to check a model for leaks instead of the two or three previously required.

  7. Vocal fold nodules in adult singers: regional opinions about etiologic factors, career impact, and treatment. A survey of otolaryngologists, speech pathologists, and teachers of singing.

    PubMed

    Hogikyan, N D; Appel, S; Guinn, L W; Haxer, M J

    1999-03-01

    This study was undertaken to better understand current regional opinions regarding vocal fold nodules in adult singers. A questionnaire was sent to 298 persons representing the 3 professional groups most involved with the care of singers with vocal nodules: otolaryngologists, speech pathologists, and teachers of singing. The questionnaire queried respondents about their level of experience with this problem, and their beliefs about causative factors, career impact, and optimum treatment. Responses within and between groups were similar, with differences between groups primarily in the magnitude of positive or negative responses, rather than in the polarity of the responses. Prevailing opinions included: recognition of causative factors in both singing and speaking voice practices, optimism about responsiveness to appropriate treatment, enthusiasm for coordinated voice therapy and voice training as first-line treatment, and acceptance of microsurgical management as appropriate treatment if behavioral management fails.

  8. Evaluation of a voice recognition system for the MOTAS pseudo pilot station function

    NASA Technical Reports Server (NTRS)

    Houck, J. A.

    1982-01-01

    The Langley Research Center has undertaken a technology development activity to provide a capability, the mission oriented terminal area simulation (MOTAS), wherein terminal area and aircraft systems studies can be performed. An experiment was conducted to evaluate state-of-the-art voice recognition technology and specifically, the Threshold 600 voice recognition system to serve as an aircraft control input device for the MOTAS pseudo pilot station function. The results of the experiment using ten subjects showed a recognition error of 3.67 percent for a 48-word vocabulary tested against a programmed vocabulary of 103 words. After the ten subjects retrained the Threshold 600 system for the words which were misrecognized or rejected, the recognition error decreased to 1.96 percent. The rejection rates for both cases were less than 0.70 percent. Based on the results of the experiment, voice recognition technology and specifically the Threshold 600 voice recognition system were chosen to fulfill this MOTAS function.

  9. Data equivalency of an interactive voice response system for home assessment of back pain and function.

    PubMed

    Shaw, William S; Verma, Santosh K

    2007-01-01

    Interactive voice response (IVR) systems that collect survey data using automated, push-button telephone responses may be useful to monitor patients' pain and function at home; however, its equivalency to other data collection methods has not been studied. To study the data equivalency of IVR measurement of pain and function to live telephone interviewing. In a prospective cohort study, 547 working adults (66% male) with acute back pain were recruited at an initial outpatient visit and completed telephone assessments one month later to track outcomes of pain, function, treatment helpfulness and return to work. An IVR system was introduced partway through the study (after the first 227 participants) to reduce the staff time necessary to contact participants by telephone during nonworking hours. Of 368 participants who were subsequently recruited and offered the IVR option, 131 (36%) used IVR, 189 (51%) were contacted by a telephone interviewer after no IVR attempt was made within five days, and 48 (13%) were lost to follow-up. Those with lower income were more likely to use IVR. Analysis of outcome measures showed that IVR respondents reported comparatively lower levels of function and less effective treatment, but not after controlling for differences due to the delay in reaching non-IVR users by telephone (mean: 35.4 versus 29.2 days). The results provided no evidence of information or selection bias associated with IVR use; however, IVR must be supplemented with other data collection options to maintain high response rates.

  10. Two experimental tests of relational models of procedural justice: non-instrumental voice and authority group membership.

    PubMed

    Platow, Michael J; Eggins, Rachael A; Chattopadhyay, Rachana; Brewer, Greg; Hardwick, Lisa; Milsom, Laurin; Brocklebank, Jacinta; Lalor, Thérèse; Martin, Rowena; Quee, Michelle; Vassallo, Sara; Welsh, Jenny

    2013-06-01

    In both a laboratory experiment (in Australia) using university as the basis of group membership, and a scenario experiment (in India) using religion as the basis of group membership, we observe more favourable respect and fairness ratings in response to an in-group authority than an out-group authority who administers non-instrumental voice. Moreover, we observe in our second experiment that reported likelihood of protest (herein called "social-change voice") was relatively high following non-instrumental voice from an out-group authority, but relatively low following non-instrumental voice from an in-group authority. Our findings are consistent with relational models of procedural justice, and extend the work by examining likely use of alternative forms of voice as well as highlighting the relative importance of instrumentality. ©2012 The British Psychological Society.

  11. Parental Engagement: Beyond Parental Involvement in Science Education

    NASA Astrophysics Data System (ADS)

    St. Louis, Kathleen

    This study critically analyzes parents' complex stories of engagement in school and science education. The purpose is not to essentialize parental involvement, but rather to understand the processes of parental involvement and push forward the current discourse on the engagement of low-income minority and immigrant parents in schools and specifically science education. Employing critical grounded theory methods over a four-year span, this study had three areas of focus. First, voices of marginalized parents in the context of various spaces within the school system are examined. Using a qualitative approach, informal, formal, and research spaces were explored along with how minority parents express voice in these various spaces. Findings indicate parents drew on capital to express voice differently in different spaces, essentially authoring new spaces or the type of engagement in existing spaces. Second, the values and beliefs of traditionally marginalized people, the Discourse of mainstream society, and how they can inform a third, more transformative space for parental engagement in science are considered. The voices of low-income, marginalized parents around science and parental engagement (i.e., first space) are contrasted with the tenets of major national science policy documents (i.e., second space). Findings indicate a disparity between the pathways of engagement for low-income parents and policymakers who shape science education. Third, methodological questions of responsibility and assumption in qualitative research are explored. The author's complex struggle to make sense of her positionality, responsibilities, and assumptions as a researcher is chronicled. Findings focused on insider/outsider issues and implications for culturally sensitive research are discussed. Finally, the implications for policy, teaching, and research are discussed.

  12. Examining Literacy Teachers' Perceptions of the Use of VoiceThread in an Elementary, Middle School, and a High School Classroom for Enhancing Instructional Goals

    ERIC Educational Resources Information Center

    Stover, Katie; Kissel, Brian; Wood, Karen; Putman, Michael

    2015-01-01

    In today's digital age, Web 2.0 tools such as VoiceThread allow users to integrate images, voices, and responses within one digital platform, providing students with the opportunity to add another layer of meaning to their texts. We conducted this research to expand our understanding of the processes necessary for integrating digital tools into…

  13. Understanding the 'Anorexic Voice' in Anorexia Nervosa.

    PubMed

    Pugh, Matthew; Waller, Glenn

    2017-05-01

    In common with individuals experiencing a number of disorders, people with anorexia nervosa report experiencing an internal 'voice'. The anorexic voice comments on the individual's eating, weight and shape and instructs the individual to restrict or compensate. However, the core characteristics of the anorexic voice are not known. This study aimed to develop a parsimonious model of the voice characteristics that are related to key features of eating disorder pathology and to determine whether patients with anorexia nervosa fall into groups with different voice experiences. The participants were 49 women with full diagnoses of anorexia nervosa. Each completed validated measures of the power and nature of their voice experience and of their responses to the voice. Different voice characteristics were associated with current body mass index, duration of disorder and eating cognitions. Two subgroups emerged, with 'weaker' and 'stronger' voice experiences. Those with stronger voices were characterized by having more negative eating attitudes, more severe compensatory behaviours, a longer duration of illness and a greater likelihood of having the binge-purge subtype of anorexia nervosa. The findings indicate that the anorexic voice is an important element of the psychopathology of anorexia nervosa. Addressing the anorexic voice might be helpful in enhancing outcomes of treatments for anorexia nervosa, but that conclusion might apply only to patients with more severe eating psychopathology. Copyright © 2016 John Wiley & Sons, Ltd. Experiences of an internal 'anorexic voice' are common in anorexia nervosa. Clinicians should consider the role of the voice when formulating eating pathology in anorexia nervosa, including how individuals perceive and relate to that voice. Addressing the voice may be beneficial, particularly in more severe and enduring forms of anorexia nervosa. When working with the voice, clinicians should aim to address both the content of the voice and how individuals relate and respond to it. Copyright © 2016 John Wiley & Sons, Ltd.

  14. Quantitative evaluation of the voice range profile in patients with voice disorder.

    PubMed

    Ikeda, Y; Masuda, T; Manako, H; Yamashita, H; Yamamoto, T; Komiyama, S

    1999-01-01

    In 1953, Calvet first displayed the fundamental frequency (pitch) and sound pressure level (intensity) of a voice on a two-dimensional plane and created a voice range profile. This profile has been used to evaluate clinically various vocal disorders, although such evaluations to date have been subjective without quantitative assessment. In the present study, a quantitative system was developed to evaluate the voice range profile utilizing a personal computer. The area of the voice range profile was defined as the voice volume. This volume was analyzed in 137 males and 175 females who were treated for various dysphonias at Kyushu University between 1984 and 1990. Ten normal subjects served as controls. The voice volume in cases with voice disorders significantly decreased irrespective of the disease and sex. Furthermore, cases having better improvement after treatment showed a tendency for the voice volume to increase. These findings illustrated the voice volume as a useful clinical test for evaluating voice control in cases with vocal disorders.

  15. A "Surprising Shock" in the Cathedral: Getting Year 7 to Vocalise Responses to the Murder of Thomas Becket

    ERIC Educational Resources Information Center

    Partridge, Mary

    2011-01-01

    Mary Partridge wanted her pupils not only to become more aware of competing and contrasting voices in the past, but to understand how historians orchestrate those voices. Using Edward Grim's eye-witness account of Thomas Becket's murder, her Year 7 pupils explored nuances in the word "shocking" as a way of distinguishing the responses of…

  16. Effects of emotional and perceptual-motor stress on a voice recognition system's accuracy: An applied investigation

    NASA Astrophysics Data System (ADS)

    Poock, G. K.; Martin, B. J.

    1984-02-01

    This was an applied investigation examining the ability of a speech recognition system to recognize speakers' inputs when the speakers were under different stress levels. Subjects were asked to speak to a voice recognition system under three conditions: (1) normal office environment, (2) emotional stress, and (3) perceptual-motor stress. Results indicate a definite relationship between voice recognition system performance and the type of low stress reference patterns used to achieve recognition.

  17. [Social consequence of a dysphonic voice, design and validation of a questionnaire and first results].

    PubMed

    Revis, J; Robieux, C; Ghio, A; Giovanni, A

    2013-01-01

    In our society, based on communication, dysphonia becomes a handicap that could be responsible of work discrimination. Actually, several commercial services are provided by phone only, and voice quality is mandatory for the employees. This work aim was to determine the social picture relayed by dysphonia. Our hypothesis was that dysphonia sounds pejorative compared to normal voice. 40 voice samples (30 dysphonic and 10 normal) were presented randomly to a perceptual jury of 20 naïve listener. The task was for each of them to fill a questionnaire, designed specifically to describe the speaker's look and personality. 20 items were evaluated, divided into 4 categories: health, temperament, appearance, and way of life. The results showed significant differences between normal subjects and dysphonic patients. For instance, the pathological voices were depicted as more tired, introverted, sloppy than normal voices, and less trustable. No significant differences were found according to the severity of voice disorders. This work is presently continued. It allowed to validate our questionnaire and has offers great perspectives on patient's management and voice therapy.

  18. The Bangor Voice Matching Test: A standardized test for the assessment of voice perception ability.

    PubMed

    Mühl, Constanze; Sheil, Orla; Jarutytė, Lina; Bestelmeyer, Patricia E G

    2017-11-09

    Recognising the identity of conspecifics is an important yet highly variable skill. Approximately 2 % of the population suffers from a socially debilitating deficit in face recognition. More recently the existence of a similar deficit in voice perception has emerged (phonagnosia). Face perception tests have been readily available for years, advancing our understanding of underlying mechanisms in face perception. In contrast, voice perception has received less attention, and the construction of standardized voice perception tests has been neglected. Here we report the construction of the first standardized test for voice perception ability. Participants make a same/different identity decision after hearing two voice samples. Item Response Theory guided item selection to ensure the test discriminates between a range of abilities. The test provides a starting point for the systematic exploration of the cognitive and neural mechanisms underlying voice perception. With a high test-retest reliability (r=.86) and short assessment duration (~10 min) this test examines individual abilities reliably and quickly and therefore also has potential for use in developmental and neuropsychological populations.

  19. Color and texture associations in voice-induced synesthesia

    PubMed Central

    Moos, Anja; Simmons, David; Simner, Julia; Smith, Rachel

    2013-01-01

    Voice-induced synesthesia, a form of synesthesia in which synesthetic perceptions are induced by the sounds of people's voices, appears to be relatively rare and has not been systematically studied. In this study we investigated the synesthetic color and visual texture perceptions experienced in response to different types of “voice quality” (e.g., nasal, whisper, falsetto). Experiences of three different groups—self-reported voice synesthetes, phoneticians, and controls—were compared using both qualitative and quantitative analysis in a study conducted online. Whilst, in the qualitative analysis, synesthetes used more color and texture terms to describe voices than either phoneticians or controls, only weak differences, and many similarities, between groups were found in the quantitative analysis. Notable consistent results between groups were the matching of higher speech fundamental frequencies with lighter and redder colors, the matching of “whispery” voices with smoke-like textures, and the matching of “harsh” and “creaky” voices with textures resembling dry cracked soil. These data are discussed in the light of current thinking about definitions and categorizations of synesthesia, especially in cases where individuals apparently have a range of different synesthetic inducers. PMID:24032023

  20. Changes after voice therapy in objective and subjective voice measurements of pediatric patients with vocal nodules.

    PubMed

    Tezcaner, Ciler Zahide; Karatayli Ozgursoy, Selmin; Ozgursoy, Selmin Karatayli; Sati, Isil; Dursun, Gursel

    2009-12-01

    The aim of this study was to analyze the efficiency of the voice therapy in children with vocal nodules by using the acoustic analysis and subjective assessment. Thirty-nine patients with vocal fold nodules, aged between 7 and 14, were included in the study. Each subject had voice therapy led by an experienced voice therapist once a week. All diagnostic and follow-up workouts were performed before the voice therapy and after the third or the sixth month. Transoral and/or transnasal videostroboscopic examination and acoustic analysis were achieved using multi-dimensional voice program (MDVP) and subjective analysis with GRBAS scale. As for the perceptual assessment, the difference was significant for four parameters out of five. A significant improvement was found in the acoustic analysis parameters of jitter, shimmer, and noise-to-harmonic ratio. The voice therapy which was planned according to patients' needs, age, compliance and response to therapy had positive effects on pediatric patients with vocal nodules. Acoustic analysis and GRBAS may be used successfully in the follow-up of pediatric vocal nodule treatment.

  1. Are 6-month-old human infants able to transfer emotional information (happy or angry) from voices to faces? An eye-tracking study.

    PubMed

    Palama, Amaya; Malsert, Jennifer; Gentaz, Edouard

    2018-01-01

    The present study examined whether 6-month-old infants could transfer amodal information (i.e. independently of sensory modalities) from emotional voices to emotional faces. Thus, sequences of successive emotional stimuli (voice or face from one sensory modality -auditory- to another sensory modality -visual-), corresponding to a cross-modal transfer, were displayed to 24 infants. Each sequence presented an emotional (angry or happy) or neutral voice, uniquely, followed by the simultaneous presentation of two static emotional faces (angry or happy, congruous or incongruous with the emotional voice). Eye movements in response to the visual stimuli were recorded with an eye-tracker. First, results suggested no difference in infants' looking time to happy or angry face after listening to the neutral voice or the angry voice. Nevertheless, after listening to the happy voice, infants looked longer at the incongruent angry face (the mouth area in particular) than the congruent happy face. These results revealed that a cross-modal transfer (from auditory to visual modalities) is possible for 6-month-old infants only after the presentation of a happy voice, suggesting that they recognize this emotion amodally.

  2. Remote voice training: A case study on space shuttle applications, appendix C

    NASA Technical Reports Server (NTRS)

    Mollakarimi, Cindy; Hamid, Tamin

    1990-01-01

    The Tile Automation System includes applications of automation and robotics technology to all aspects of the Shuttle tile processing and inspection system. An integrated set of rapid prototyping testbeds was developed which include speech recognition and synthesis, laser imaging systems, distributed Ada programming environments, distributed relational data base architectures, distributed computer network architectures, multi-media workbenches, and human factors considerations. Remote voice training in the Tile Automation System is discussed. The user is prompted over a headset by synthesized speech for the training sequences. The voice recognition units and the voice output units are remote from the user and are connected by Ethernet to the main computer system. A supervisory channel is used to monitor the training sequences. Discussions include the training approaches as well as the human factors problems and solutions for this system utilizing remote training techniques.

  3. Speaking in Character: Voice Communication in Virtual Worlds

    NASA Astrophysics Data System (ADS)

    Wadley, Greg; Gibbs, Martin R.

    This chapter summarizes 5 years of research on the implications of introducing voice communication systems to virtual worlds. Voice introduces both benefits and problems for players of fast-paced team games, from better coordination of groups and greater social presence of fellow players on the positive side, to negative features such as channel congestion, transmission of noise, and an unwillingness by some to use voice with strangers online. Similarly, in non-game worlds like Second Life, issues related to identity and impression management play important roles, as voice may build greater trust that is especially important for business users, yet it erodes the anonymity and ability to conceal social attributes like gender that are important for other users. A very different mixture of problems and opportunities exists when users conduct several simultaneous conversations in multiple text and voice channels. Technical difficulties still exist with current systems, including the challenge of debugging and harmonizing all the participants' voice setups. Different groups use virtual worlds for very different purposes, so a single modality may not suit all.

  4. Building VoiceXML-Based Applications

    DTIC Science & Technology

    2002-01-01

    basketball games. The Busline systems were pri- y developed using an early implementation of VoiceXML he NBA Update Line was developed using VoiceXML...traveling in and out of Pittsburgh’s rsity neighborhood. The second project is the NBA Up- Line, which provides callers with real-time information NBA ... NBA UPDATE LINE The target user of this system is a fairly knowledgeable basket- ball fan; the system must therefore be able to provide detailed

  5. DTO-675: Voice Control of the Closed Circuit Television System

    NASA Technical Reports Server (NTRS)

    Salazar, George; Gaston, Darilyn M.; Haynes, Dena S.

    1996-01-01

    This report presents the results of the Detail Test Object (DTO)-675 "Voice Control of the Closed Circuit Television (CCTV)" system. The DTO is a follow-on flight of the Voice Command System (VCS) that flew as a secondary payload on STS-41. Several design changes were made to the VCS for the STS-78 mission. This report discusses those design changes, the data collected during the mission, recognition problems encountered, and findings.

  6. Voice stress analysis and evaluation

    NASA Astrophysics Data System (ADS)

    Haddad, Darren M.; Ratley, Roy J.

    2001-02-01

    Voice Stress Analysis (VSA) systems are marketed as computer-based systems capable of measuring stress in a person's voice as an indicator of deception. They are advertised as being less expensive, easier to use, less invasive in use, and less constrained in their operation then polygraph technology. The National Institute of Justice have asked the Air Force Research Laboratory for assistance in evaluating voice stress analysis technology. Law enforcement officials have also been asking questions about this technology. If VSA technology proves to be effective, its value for military and law enforcement application is tremendous.

  7. The distress of voice-hearing: the use of simulation for awareness, understanding and communication skill development in undergraduate nursing education.

    PubMed

    Orr, Fiona; Kellehear, Kevin; Armari, Elizabeth; Pearson, Arana; Holmes, Douglas

    2013-11-01

    Role-play scenarios are frequently used with undergraduate nursing students enrolled in mental health nursing subjects to simulate the experience of voice-hearing. However, role-play has limitations and typically does not involve those who hear voices. This collaborative project between mental health consumers who hear voices and nursing academics aimed to develop and assess simulated voice-hearing as an alternative learning tool that could provide a deeper understanding of the impact of voice-hearing, whilst enabling students to consider the communication skills required when interacting with voice-hearers. Simulated sounds and voices recorded by consumers on mp3 players were given to eighty final year nursing students undertaking a mental health elective. Students participated in various activities whilst listening to the simulations. Seventy-six (95%) students completed a written evaluation following the simulation, which assessed the benefits of the simulation and its implications for clinical practice. An analysis of the students' responses by an external evaluator indicated that there were three major learning outcomes: developing an understanding of voice-hearing, increasing students' awareness of its impact on functioning, and consideration of the communication skills necessary to engage with consumers who hear voices. Copyright © 2013 Elsevier Ltd. All rights reserved.

  8. China-U.S. Relations: Current Issues and Implications for U.S. Policy

    DTIC Science & Technology

    2008-10-09

    United States willing to work cooperatively with China while encouraging Beijing to become a “responsible stakeholder” in the global system. U.S. Treasury...Secretary Henry Paulson in December 2006 established a U.S.-China Strategic Economic Dialogue with Beijing , the most senior regular dialogue yet held...beyond. Beijing’s response has led some Tibetan activists to add their voices to other calls urging a boycott of the Summer Olympics in Beijing in

  9. Automatic translation among spoken languages

    NASA Technical Reports Server (NTRS)

    Walter, Sharon M.; Costigan, Kelly

    1994-01-01

    The Machine Aided Voice Translation (MAVT) system was developed in response to the shortage of experienced military field interrogators with both foreign language proficiency and interrogation skills. Combining speech recognition, machine translation, and speech generation technologies, the MAVT accepts an interrogator's spoken English question and translates it into spoken Spanish. The spoken Spanish response of the potential informant can then be translated into spoken English. Potential military and civilian applications for automatic spoken language translation technology are discussed in this paper.

  10. Speaking more broadly: an examination of the nature, antecedents, and consequences of an expanded set of employee voice behaviors.

    PubMed

    Maynes, Timothy D; Podsakoff, Philip M

    2014-01-01

    Scholarly interest in employee voice behavior has increased dramatically over the past 15 years. Although this research has produced valuable knowledge, it has focused almost exclusively on voice as a positively intended challenge to the status quo, even though some scholars have argued that it need not challenge the status quo or be well intentioned. Thus, in this paper, we create an expanded view of voice; one that extends beyond voice as a positively intended challenge to the status quo to include voice that supports how things are being done in organizations as well as voice that may not be well intentioned. We construct a framework based on this expanded view that identifies 4 different types of voice behavior (supportive, constructive, defensive, and destructive). We then develop and validate survey measures for each of these. Evidence from 5 studies across 4 samples provides strong support for our new measures in that (a) a 4-factor confirmatory factor analysis model fit the data significantly better than 1-, 2-, or 3-factor models; (b) the voice measures converged with and yet remained distinct from conceptually related comparison constructs; (c) personality predictors exhibited unique patterns of relationships with the different types of voice; (d) variations in actual voice behaviors had a direct causal impact on responses to the survey items; and (e) each type of voice significantly impacted important outcomes for voicing employees (e.g., likelihood of relying on a voicing employee's opinions and evaluations of a voicing employee's overall performance). Implications of our findings are discussed. PsycINFO Database Record (c) 2014 APA, all rights reserved

  11. Validation and Adaptation of the Singing Voice Handicap Index for Egyptian Singing Voice.

    PubMed

    Abou-Elsaad, Tamer; Baz, Hemmat; Afsah, Omayma; Abo-Elsoud, Hend

    2017-01-01

    Measuring the severity of a voice disorder is difficult. This can be achieved by both subjective and objective measures. The Voice Handicap Index is the most known and used self-rating tool for voice disorders. The Classical Singing Handicap Index (CSHI) is a self-administered questionnaire measuring the impact of vocal deviation on the quality of life of singers. The objective of this study was to develop an Arabic version of the CSHI and to test its validity and reliability in Egyptian singers with different singing styles with normal voice and with voice disorders. The interpreted version was administered to 70 Egyptian singers including artistic singers (classical and popular) and specialized singers (Quran reciters and priests) who were divided into 40 asymptomatic singers (control group) and 30 singers with voice disorders. Participants' responses were statistically analyzed to assess the validity and reliability, and to compare the patient group with the control group. Quran reciters, patients with no previous professional training, and patients with vocal fold lesions demonstrated the highest scores. The Arabic version of CSHI is found to be a reliable, valid, and sensitive self-assessment tool that can be used in the clinical practice for the evaluation of the impact of voice disorders on singing voice. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  12. Intra- and Inter-database Study for Arabic, English, and German Databases: Do Conventional Speech Features Detect Voice Pathology?

    PubMed

    Ali, Zulfiqar; Alsulaiman, Mansour; Muhammad, Ghulam; Elamvazuthi, Irraivan; Al-Nasheri, Ahmed; Mesallam, Tamer A; Farahat, Mohamed; Malki, Khalid H

    2017-05-01

    A large population around the world has voice complications. Various approaches for subjective and objective evaluations have been suggested in the literature. The subjective approach strongly depends on the experience and area of expertise of a clinician, and human error cannot be neglected. On the other hand, the objective or automatic approach is noninvasive. Automatic developed systems can provide complementary information that may be helpful for a clinician in the early screening of a voice disorder. At the same time, automatic systems can be deployed in remote areas where a general practitioner can use them and may refer the patient to a specialist to avoid complications that may be life threatening. Many automatic systems for disorder detection have been developed by applying different types of conventional speech features such as the linear prediction coefficients, linear prediction cepstral coefficients, and Mel-frequency cepstral coefficients (MFCCs). This study aims to ascertain whether conventional speech features detect voice pathology reliably, and whether they can be correlated with voice quality. To investigate this, an automatic detection system based on MFCC was developed, and three different voice disorder databases were used in this study. The experimental results suggest that the accuracy of the MFCC-based system varies from database to database. The detection rate for the intra-database ranges from 72% to 95%, and that for the inter-database is from 47% to 82%. The results conclude that conventional speech features are not correlated with voice, and hence are not reliable in pathology detection. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  13. Victims' Voices in the Correctional Setting: Cognitive Gains in an Offender Education Program

    ERIC Educational Resources Information Center

    Monahan, Lynn Hunt; Monahan, James J.; Gaboury, Mario T.; Niesyn, Patricia A.

    2004-01-01

    A 40-hour education program covering the impact of crime on victims and their families was offered to sentenced offenders in a statewide prison system. Questionnaire responses from 339 males, ranging in age from 21-45 years, were examined. Results indicated that the program increased knowledge of victim rights, facts of victimization, and…

  14. Design and control of six degree-of-freedom active vibration isolation table.

    PubMed

    Hong, Jinpyo; Park, Kyihwan

    2010-03-01

    A six-axis active vibration isolation system (AVIS) is designed by using the direct driven guide and ball contact mechanisms in order to have no cross-coupling between actuators. The point contact configuration gives an advantage of having an easy assembly of eight voice coil actuators to an upper and a base plate. A voice coil actuator is used since it can provide a large displacement and sufficient bandwidth required for vibration control. The AVIS is controlled considering the effect of flexible vibration mode in the upper plate and velocity sensor dynamics. A loop shaping technique and phase margin condition are applied to design a vibration controller. The performances of the AVIS are investigated in the frequency domain and finally validated by comparing with the passive isolation system. The scanning profiles of the specimen are compared together by using the atomic force microscope. The robustness of the AVIS is verified by showing the impulse response.

  15. Design and control of six degree-of-freedom active vibration isolation table

    NASA Astrophysics Data System (ADS)

    Hong, Jinpyo; Park, Kyihwan

    2010-03-01

    A six-axis active vibration isolation system (AVIS) is designed by using the direct driven guide and ball contact mechanisms in order to have no cross-coupling between actuators. The point contact configuration gives an advantage of having an easy assembly of eight voice coil actuators to an upper and a base plate. A voice coil actuator is used since it can provide a large displacement and sufficient bandwidth required for vibration control. The AVIS is controlled considering the effect of flexible vibration mode in the upper plate and velocity sensor dynamics. A loop shaping technique and phase margin condition are applied to design a vibration controller. The performances of the AVIS are investigated in the frequency domain and finally validated by comparing with the passive isolation system. The scanning profiles of the specimen are compared together by using the atomic force microscope. The robustness of the AVIS is verified by showing the impulse response.

  16. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2002-01-01

    Low power EM waves are used to detect motions of vocal tract tissues of the human speech system before, during, and after voiced speech. A voiced excitation function is derived. The excitation function provides speech production information to enhance speech characterization and to enable noise removal from human speech.

  17. Performance of wavelet analysis and neural networks for pathological voices identification

    NASA Astrophysics Data System (ADS)

    Salhi, Lotfi; Talbi, Mourad; Abid, Sabeur; Cherif, Adnane

    2011-09-01

    Within the medical environment, diverse techniques exist to assess the state of the voice of the patient. The inspection technique is inconvenient for a number of reasons, such as its high cost, the duration of the inspection, and above all, the fact that it is an invasive technique. This study focuses on a robust, rapid and accurate system for automatic identification of pathological voices. This system employs non-invasive, non-expensive and fully automated method based on hybrid approach: wavelet transform analysis and neural network classifier. First, we present the results obtained in our previous study while using classic feature parameters. These results allow visual identification of pathological voices. Second, quantified parameters drifting from the wavelet analysis are proposed to characterise the speech sample. On the other hand, a system of multilayer neural networks (MNNs) has been developed which carries out the automatic detection of pathological voices. The developed method was evaluated using voice database composed of recorded voice samples (continuous speech) from normophonic or dysphonic speakers. The dysphonic speakers were patients of a National Hospital 'RABTA' of Tunis Tunisia and a University Hospital in Brussels, Belgium. Experimental results indicate a success rate ranging between 75% and 98.61% for discrimination of normal and pathological voices using the proposed parameters and neural network classifier. We also compared the average classification rate based on the MNN, Gaussian mixture model and support vector machines.

  18. A comparison of the VHI, VHI-10, and V-RQOL for measuring the effect of botox therapy in adductor spasmodic dysphonia.

    PubMed

    Morzaria, Sanjay; Damrose, Edward J

    2012-05-01

    Although disease-specific quality-of-life (QOL) instruments are an invaluable outcome measure in spasmodic dysphonia, there is no consensus on which QOL instrument should be used. To determine the responsiveness of the Voice Handicap Index (VHI), Voice Handicap Index-10 (VHI-10), and Voice-Related Quality of Life (V-RQOL) to the treatment effect of botulinum toxin (Botox) in adductor spasmodic dysphonia (ADSD). Stanford University Voice and Swallowing Center. Prospective case series (level of evidence=4). Consecutive ADSD patients with a stable Botox dose-response relationship were recruited prospectively. VHI, VHI-10, and V-RQOL scores were obtained pretreatment and during the middle third of the posttreatment injection cycle. Thrity-seven patients completed the follow-up. The average total Botox dose was 0.88 units. The average follow-up time after injection was 7.84 weeks. The pretreatment QOL scores reflected the burden of the disease. All the three instruments were highly correlated in subscale and total scores. After treatment, all three instruments showed significant improvement. The VHI, VHI-10, and V-RQOL all reflected the morbidity associated with ADSD and were significantly responsive to the effect of Botox therapy. The choice of instrument should be based on physician preference. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  19. [The comparative assessment of the vocal function in the professional voice users and non-occupational voice users in the late adulthood].

    PubMed

    Pavlikhin, O G; Romanenko, S G; Krasnikova, D I; Lesogorova, E V; Yakovlev, V S

    The objective of the present study was to evaluate the clinical and functional condition of the voice apparatus in the elderly patients and to elaborate recommendations for the prevention of disturbances of the vocal function in the professional voice users. This comprehensive study involved 95 patients including the active professional voice users (n=48) and 45 non-occupational voice users at the age from 61 to 82 years with the employment history varying from 32 to 51 years. The study was designed to obtain the voice characteristics by means of the subjective auditory assessment, microlaryngoscopy, video laryngostroboscopy, determination of maximum phonation time (MPT), and computer-assisted acoustic analysis of the voice with the use of the MDVP Kay Pentaxy system. The level of anxiety of the patients was estimated based on the results of the HADS questionnaire study. It is concluded that the majority of the disturbances of the vocal function in the professional voice users have the functional nature. It is concluded that the method of neuro-muscular electrophonopedic stimulation (NMEPS) of laryngeal muscles is the method of choice for the diagnostics of the vocal function of the voice users in the late adulthood. It is recommended that the professional vocal load for such subjects should not exceed 12-14 hours per week. Rational psychotherapy must constitute an important component of the system of measures intended to support the working capacity of the voice users belonging to this age group.

  20. QM/PSK Voice/Data Modem

    DOT National Transportation Integrated Search

    1976-03-01

    Two Quadrature Modulation/Phase Shift Keyed (QM/PSK) Voice/Data Modem systems have been developed as part of the satellite communications hardware for advanced air traffic control systems. These systems consist of a modulator and demodulator unti whi...

  1. Voice loops as coordination aids in space shuttle mission control.

    PubMed

    Patterson, E S; Watts-Perotti, J; Woods, D D

    1999-01-01

    Voice loops, an auditory groupware technology, are essential coordination support tools for experienced practitioners in domains such as air traffic management, aircraft carrier operations and space shuttle mission control. They support synchronous communication on multiple channels among groups of people who are spatially distributed. In this paper, we suggest reasons for why the voice loop system is a successful medium for supporting coordination in space shuttle mission control based on over 130 hours of direct observation. Voice loops allow practitioners to listen in on relevant communications without disrupting their own activities or the activities of others. In addition, the voice loop system is structured around the mission control organization, and therefore directly supports the demands of the domain. By understanding how voice loops meet the particular demands of the mission control environment, insight can be gained for the design of groupware tools to support cooperative activity in other event-driven domains.

  2. A survey of the state-of-the-art and focused research in range systems, task 1

    NASA Technical Reports Server (NTRS)

    Omura, J. K.

    1986-01-01

    This final report presents the latest research activity in voice compression. We have designed a non-real time simulation system that is implemented around the IBM-PC where the IBM-PC is used as a speech work station for data acquisition and analysis of voice samples. A real-time implementation is also proposed. This real-time Voice Compression Board (VCB) is built around the Texas Instruments TMS-3220. The voice compression algorithm investigated here was described in an earlier report titled, Low Cost Voice Compression for Mobile Digital Radios, by the author. We will assume the reader is familiar with the voice compression algorithm discussed in this report. The VCB compresses speech waveforms at data rates ranging from 4.8 K bps to 16 K bps. This board interfaces to the IBM-PC 8-bit bus, and plugs into a single expansion slot on the mother board.

  3. Voice loops as coordination aids in space shuttle mission control

    NASA Technical Reports Server (NTRS)

    Patterson, E. S.; Watts-Perotti, J.; Woods, D. D.

    1999-01-01

    Voice loops, an auditory groupware technology, are essential coordination support tools for experienced practitioners in domains such as air traffic management, aircraft carrier operations and space shuttle mission control. They support synchronous communication on multiple channels among groups of people who are spatially distributed. In this paper, we suggest reasons for why the voice loop system is a successful medium for supporting coordination in space shuttle mission control based on over 130 hours of direct observation. Voice loops allow practitioners to listen in on relevant communications without disrupting their own activities or the activities of others. In addition, the voice loop system is structured around the mission control organization, and therefore directly supports the demands of the domain. By understanding how voice loops meet the particular demands of the mission control environment, insight can be gained for the design of groupware tools to support cooperative activity in other event-driven domains.

  4. Analysis of Jordan’s Proposed Emergency Communication Interoperability Plan (JECIP) for Disaster Response

    DTIC Science & Technology

    2008-12-01

    Transmission quality measurements start once the call is established which includes low voice volume, level of noise , echo, crosstalk, and garbling...to failure, and finally, there is restorability which is a measure of how easy the system is restored upon failure. To reduce frequency of failure...Silicon and Germanium. These systems are friendly for the environment, have low - noise , have no fuel consumption, are maintenance-free, and have no

  5. Automated conversation system before pediatric primary care visits: a randomized trial.

    PubMed

    Adams, William G; Phillips, Barrett D; Bacic, Janine D; Walsh, Kathleen E; Shanahan, Christopher W; Paasche-Orlow, Michael K

    2014-09-01

    Interactive voice response systems integrated with electronic health records have the potential to improve primary care by engaging parents outside clinical settings via spoken language. The objective of this study was to determine whether use of an interactive voice response system, the Personal Health Partner (PHP), before routine health care maintenance visits could improve the quality of primary care visits and be well accepted by parents and clinicians. English-speaking parents of children aged 4 months to 11 years called PHP before routine visits and were randomly assigned to groups by the system at the time of the call. Parents' spoken responses were used to provide tailored counseling and support goal setting for the upcoming visit. Data were transferred to the electronic health records for review during visits. The study occurred in an urban hospital-based pediatric primary care center. Participants were called after the visit to assess (1) comprehensiveness of screening and counseling, (2) assessment of medications and their management, and (3) parent and clinician satisfaction. PHP was able to identify and counsel in multiple areas. A total of 9.7% of parents responded to the mailed invitation. Intervention parents were more likely to report discussing important issues such as depression (42.6% vs 25.4%; P < .01) and prescription medication use (85.7% vs 72.6%; P = .04) and to report being better prepared for visits. One hundred percent of clinicians reported that PHP improved the quality of their care. Systems like PHP have the potential to improve clinical screening, counseling, and medication management. Copyright © 2014 by the American Academy of Pediatrics.

  6. Human voice quality measurement in noisy environments.

    PubMed

    Ueng, Shyh-Kuang; Luo, Cheng-Ming; Tsai, Tsung-Yu; Yeh, Hsuan-Chen

    2015-01-01

    Computerized acoustic voice measurement is essential for the diagnosis of vocal pathologies. Previous studies showed that ambient noises have significant influences on the accuracy of voice quality assessment. This paper presents a voice quality assessment system that can accurately measure qualities of voice signals, even though the input voice data are contaminated by low-frequency noises. The ambient noises in our living rooms and laboratories are collected and the frequencies of these noises are analyzed. Based on the analysis, a filter is designed to reduce noise level of the input voice signal. Then, improved numerical algorithms are employed to extract voice parameters from the voice signal to reveal the health of the voice signal. Compared with MDVP and Praat, the proposed method outperforms these two widely used programs in measuring fundamental frequency and harmonic-to-noise ratio, and its performance is comparable to these two famous programs in computing jitter and shimmer. The proposed voice quality assessment method is resistant to low-frequency noises and it can measure human voice quality in environments filled with noises from air-conditioners, ceiling fans and cooling fans of computers.

  7. Effects of Voice Coding and Speech Rate on a Synthetic Speech Display in a Telephone Information System

    DTIC Science & Technology

    1988-05-01

    Seeciv Limited- System for varying Senses term filter capacity output until some Figure 2. Original limited-capacity channel model (Frim Broadbent, 1958) S...2 Figure 2. Original limited-capacity channel model (From Broadbent, 1958) .... 10 Figure 3. Experimental...unlimited variety of human voices for digital recording sources. Synthesis by Analysis Analysis-synthesis methods electronically model the human voice

  8. Writing with Voice: An Investigation of the Use of a Voice Recognition System as a Writing Aid for a Man with Aphasia

    ERIC Educational Resources Information Center

    Bruce, Carolyn; Edmundson, Anne; Coleman, Michael

    2003-01-01

    Background: People with aphasia may experience difficulties that prevent them from demonstrating in writing what they know and can produce orally. Voice recognition systems that allow the user to speak into a microphone and see their words appear on a computer screen have the potential to assist written communication. Aim: This study investigated…

  9. A General Purpose Connections type CTI Server Based on SIP Protocol and Its Implementation

    NASA Astrophysics Data System (ADS)

    Watanabe, Toru; Koizumi, Hisao

    In this paper, we propose a general purpose connections type CTI (Computer Telephony Integration) server that provides various CTI services such as voice logging where the CTI server communicates with IP-PBX using the SIP (Session Initiation Protocol), and accumulates voice packets of external line telephone call flowing between an IP telephone for extension and a VoIP gateway connected to outside line networks. The CTI server realizes CTI services such as voice logging, telephone conference, or IVR (interactive voice response) with accumulating and processing voice packets sampled. Furthermore, the CTI server incorporates a web server function which can provide various CTI services such as a Web telephone directory via a Web browser to PCs, cellular telephones or smart-phones in mobile environments.

  10. Monitoring daily affective symptoms and memory function using interactive voice response in outpatients receiving electroconvulsive therapy.

    PubMed

    Fazzino, Tera L; Rabinowitz, Terry; Althoff, Robert R; Helzer, John E

    2013-12-01

    Recently, there has been a gradual shift from inpatient-only electroconvulsive therapy (ECT) toward outpatient administration. Potential advantages include convenience and reduced cost. But providers do not have the same opportunity to monitor treatment response and adverse effects as they do with inpatients. This can obviate some of the potential advantages of outpatient ECT, such as tailoring treatment intervals to clinical response. Scheduling is typically algorithmic rather than empirically based. Daily monitoring through an automated telephone, interactive voice response (IVR), is a potential solution to this quandary. To test feasibility of clinical monitoring via IVR, we recruited 26 patients (69% female; mean age, 51 years) receiving outpatient ECT to make daily IVR reports of affective symptoms and subjective memory for 60 days. The IVR also administered a word recognition task daily to test objective memory. Every seventh day, a longer IVR weekly interview included questions about suicidal ideation. Overall daily call compliance was high (mean, 80%). Most participants (96%) did not consider the calls to be time-consuming. Longitudinal regression analysis using generalized estimating equations revealed that participant objective memory functioning significantly improved during the study (P < 0.05). Of 123 weekly IVR interviews, 41 reports (33%) in 14 patients endorsed suicidal ideation during the previous week. Interactive voice response monitoring of outpatient ECT can provide more detailed clinical information than standard outpatient ECT assessment. Interactive voice response data offer providers a comprehensive, longitudinal picture of patient treatment response and adverse effects as a basis for treatment scheduling and ongoing clinical management.

  11. Dissociation and psychosis in dissociative identity disorder and schizophrenia.

    PubMed

    Laddis, Andreas; Dell, Paul F

    2012-01-01

    Dissociative symptoms, first-rank symptoms of schizophrenia, and delusions were assessed in 40 schizophrenia patients and 40 dissociative identity disorder (DID) patients with the Multidimensional Inventory of Dissociation (MID). Schizophrenia patients were diagnosed with the Structured Clinical Interview for the DSM-IV Axis I Disorders; DID patients were diagnosed with the Structured Clinical Interview for DSM-IV Dissociative Disorders-Revised. DID patients obtained significantly (a) higher dissociation scores; (b) higher passive-influence scores (first-rank symptoms); and (c) higher scores on scales that measure child voices, angry voices, persecutory voices, voices arguing, and voices commenting. Schizophrenia patients obtained significantly higher delusion scores than did DID patients. What is odd is that the dissociation scores of schizophrenia patients were unrelated to their reports of childhood maltreatment. Multiple regression analyses indicated that 81% of the variance in DID patients' dissociation scores was predicted by the MID's Ego-Alien Experiences Scale, whereas 92% of the variance in schizophrenia patients' dissociation scores was predicted by the MID's Voices Scale. We propose that schizophrenia patients' responses to the MID do not index the same pathology as do the responses of DID patients. We argue that neither phenomenological definitions of dissociation nor the current generation of dissociation instruments (which are uniformly phenomenological in nature) can distinguish between the dissociative phenomena of DID and what we suspect are just the dissociation-like phenomena of schizophrenia.

  12. A simulation study of the effects of communication delay on air traffic control

    DOT National Transportation Integrated Search

    1990-09-01

    This study was conducted to examine the impacts of voice communications delays : characteristic of Voice Switching and Control System (VSCS) and satellite : communications systems on air traffic system performance, controller stress : and workload, a...

  13. Decoding Articulatory Features from fMRI Responses in Dorsal Speech Regions.

    PubMed

    Correia, Joao M; Jansma, Bernadette M B; Bonte, Milene

    2015-11-11

    The brain's circuitry for perceiving and producing speech may show a notable level of overlap that is crucial for normal development and behavior. The extent to which sensorimotor integration plays a role in speech perception remains highly controversial, however. Methodological constraints related to experimental designs and analysis methods have so far prevented the disentanglement of neural responses to acoustic versus articulatory speech features. Using a passive listening paradigm and multivariate decoding of single-trial fMRI responses to spoken syllables, we investigated brain-based generalization of articulatory features (place and manner of articulation, and voicing) beyond their acoustic (surface) form in adult human listeners. For example, we trained a classifier to discriminate place of articulation within stop syllables (e.g., /pa/ vs /ta/) and tested whether this training generalizes to fricatives (e.g., /fa/ vs /sa/). This novel approach revealed generalization of place and manner of articulation at multiple cortical levels within the dorsal auditory pathway, including auditory, sensorimotor, motor, and somatosensory regions, suggesting the representation of sensorimotor information. Additionally, generalization of voicing included the right anterior superior temporal sulcus associated with the perception of human voices as well as somatosensory regions bilaterally. Our findings highlight the close connection between brain systems for speech perception and production, and in particular, indicate the availability of articulatory codes during passive speech perception. Sensorimotor integration is central to verbal communication and provides a link between auditory signals of speech perception and motor programs of speech production. It remains highly controversial, however, to what extent the brain's speech perception system actively uses articulatory (motor), in addition to acoustic/phonetic, representations. In this study, we examine the role of articulatory representations during passive listening using carefully controlled stimuli (spoken syllables) in combination with multivariate fMRI decoding. Our approach enabled us to disentangle brain responses to acoustic and articulatory speech properties. In particular, it revealed articulatory-specific brain responses of speech at multiple cortical levels, including auditory, sensorimotor, and motor regions, suggesting the representation of sensorimotor information during passive speech perception. Copyright © 2015 the authors 0270-6474/15/3515015-11$15.00/0.

  14. Assessment of an interactive voice response system for identifying falls in a statewide sample of older adults.

    PubMed

    Albert, Steven M; King, Jennifer; Keene, Robert M

    2015-02-01

    Interactive voice response (IVR) systems offer great advantages for data collection in large, geographically dispersed samples involving frequent contact. We assessed the quality of IVR data collected from older respondents participating in a statewide falls prevention program evaluation in Pennsylvania in 2010-12. Participants (n=1834) were followed up monthly for up to 10 months to compare respondents who completed all, some, or no assessments in the IVR system. Validity was assessed by examining IVR-reported falls incidence relative to baseline in-person self-report and performance assessment of balance. While a third of the sample switched from IVR to in-person calls over follow-up, IVR interviews were successfully used to complete 68.1% of completed monthly assessments (10,511/15,430). Switching to in-person interviews was not associated with measures of participant function or cognition. Both self-reported (p<.0001) and performance assessment of balance (p=.05) at baseline were related to falls incidence. IVR is a productive modality for falls research among older adults. Future research should establish what level of initial personal research contact is optimal for boosting IVR completion rates and what research domains are most appropriate for this kind of contact. Copyright © 2014 Elsevier Inc. All rights reserved.

  15. Do What I Say! Voice Recognition Makes Major Advances.

    ERIC Educational Resources Information Center

    Ruley, C. Dorsey

    1994-01-01

    Explains voice recognition technology applications in the workplace, schools, and libraries. Highlights include a voice-controlled work station using the DragonDictate system that can be used with dyslexic students, converting text to speech, and converting speech to text. (LRW)

  16. Spanish-Speaking Patients’ Engagement in Interactive Voice Response (IVR) Chronic Disease Self-Management Support Calls: Analyses of Data from Three Countries

    PubMed Central

    Piette, John D.; Marinec, Nicolle; Gallegos-Cabriales, Esther C.; Gutierrez-Valverde, Juana Mercedes; Rodriguez-Saldaña, Joel; Mendoz-Alevares, Milton; Silveira, Maria J.

    2013-01-01

    We used data from Interactive Voice Response (IVR) self-management support studies in Honduras, Mexico, and the United States (US) to determine whether IVR calls to Spanish-speaking patients with chronic illnesses is a feasible strategy for improving monitoring and education between face-to-face visits. 268 patients with diabetes or hypertension participated in 6–12 weeks of weekly IVR follow-up. IVR calls emanated from US servers with connections via Voice over IP. More than half (54%) of patients enrolled with an informal caregiver who received automated feedback based on the patient’s assessments, and clinical staff received urgent alerts. Participants had on average 6.1 years of education, and 73% were women. After 2,443 person weeks of follow-up, patients completed 1,494 IVR assessments. Call completion rates were higher in the US (75%) than in Honduras (59%) or Mexico (61%; p<0.001). Patients participating with an informal caregiver were more likely to complete calls (adjusted odds ratio [AOR]: 1.53; 95% confidence interval [CI]: 1.04, 2.25) while patients reporting fair or poor health at enrollment were less likely (AOR:0.59; 95% CI: 0.38, 0.92). Satisfaction rates were high, with 98% of patients reporting that the system was easy to use, and 86% reporting that the calls helped them a great deal in managing their health problems. In summary, IVR self-management support is feasible among Spanish-speaking patients with chronic disease, including those living in less-developed countries. Voice over IP can be used to deliver IVR disease management services internationally; involving informal caregivers may increase patient engagement. PMID:23532005

  17. Powerlessness, marginalized identity, and silencing of health concerns: voiced realities of women living with a mental health diagnosis.

    PubMed

    Van Den Tillaart, Susan; Kurtz, Donna; Cash, Penny

    2009-06-01

    Using a feminist qualitative approach, this study substantiated many earlier research findings that document how women with a mental health diagnosis experience unequal access to comprehensive health care compared to the general population. Accounts of this disparity are documented in the literature, yet the literature has failed to record or attend to the voices of those living with mental health challenges. In this paper, women living with a mental health diagnosis describe their experiences as they interface with the health-care system. The participating women's stories clearly relate the organizational and interpersonal challenges commonly faced when they seek health-care services. The stories include experiences of marginalized identity, powerlessness, and silencing of voiced health concerns. The women tell of encountered gaps in access to health care and incomplete health assessment, screening, and treatment. It becomes clear that personal and societal stigmatization related to the mental health diagnosis plays a significant role in these isolating and unsatisfactory experiences. Lastly, the women offer beginning ideas for change by suggesting starting points to eliminate the institutional and interpersonal obstacles or barriers to their wellness. The concerns raised demand attention, reconsideration, and change by those in the health-care system responsible for policy and practice.

  18. The Impacts of the Voice Change, Grade Level, and Experience on the Singing Self-Efficacy of Emerging Adolescent Males

    ERIC Educational Resources Information Center

    Fisher, Ryan A.

    2014-01-01

    The purposes of the study are to describe characteristics of the voice change in sixth-, seventh-, and eighth-grade choir students using Cooksey's voice-change classification system and to determine if the singing self-efficacy of adolescent males is affected by the voice change, grade level, and experience. Participants (N = 80) consisted of…

  19. Voices for Diversity.

    ERIC Educational Resources Information Center

    Future Teacher, 1995

    1995-01-01

    Prominent Americans were asked to reflect on the diversity challenge facing America's teacher workforce. The following leaders from several fields voiced their support of teachers and their beliefs America needs more diverse and culturally responsive teachers: (1) Mary Hatwood Futrell, President of Education International; (2) Carol Moseley-Braun,…

  20. A mobile communication system providing integrated voice/data services over power limited satellite channels

    NASA Astrophysics Data System (ADS)

    Bose, Sanjay K.; Gordon, J. J.

    The modeling and analysis of a system providing integrated voice/data services to mobile terminals over a power-limited satellite channel are discussed. The mobiles use slotted Aloha random access to send requests for channel assignments to a central station. For successful requests, the actual transmission of voice/data within a call is done using the channel assigned for this purpose by the central station. The satellite channel is assumed to be power limited. Taking into account the known burstiness of voice sources (which use a voice-activated switch), the central station overassigns channels so that the average total power is below the power limit of the satellite transponder. The performance of this model is analyzed. Certain simple, static control strategies for improving performance are also proposed.

  1. Voice and gesture-based 3D multimedia presentation tool

    NASA Astrophysics Data System (ADS)

    Fukutake, Hiromichi; Akazawa, Yoshiaki; Okada, Yoshihiro

    2007-09-01

    This paper proposes a 3D multimedia presentation tool that allows the user to manipulate intuitively only through the voice input and the gesture input without using a standard keyboard or a mouse device. The authors developed this system as a presentation tool to be used in a presentation room equipped a large screen like an exhibition room in a museum because, in such a presentation environment, it is better to use voice commands and the gesture pointing input rather than using a keyboard or a mouse device. This system was developed using IntelligentBox, which is a component-based 3D graphics software development system. IntelligentBox has already provided various types of 3D visible, reactive functional components called boxes, e.g., a voice input component and various multimedia handling components. IntelligentBox also provides a dynamic data linkage mechanism called slot-connection that allows the user to develop 3D graphics applications by combining already existing boxes through direct manipulations on a computer screen. Using IntelligentBox, the 3D multimedia presentation tool proposed in this paper was also developed as combined components only through direct manipulations on a computer screen. The authors have already proposed a 3D multimedia presentation tool using a stage metaphor and its voice input interface. This time, we extended the system to make it accept the user gesture input besides voice commands. This paper explains details of the proposed 3D multimedia presentation tool and especially describes its component-based voice and gesture input interfaces.

  2. An efficient protocol for providing integrated voice/data services to mobiles over power-limited satellite channels

    NASA Astrophysics Data System (ADS)

    Bose, Sanjay K.

    1991-02-01

    Various mobile satellite communication systems are being developed for providing integrated voice/data services over a shared satellite transponder which is power-limited in nature. A common strategy is to use slotted ALOHA request channels to request channel assignments for voice/data calls from a network management station. To maximize efficiency in a system with a power-limited satellite transponder, it is proposed that the bursty nature of voice sources be exploited by the NMS to 'over-assign' channels. This may cause problems of inefficiency and potential instability, as well as a degradation in the quality of service. Augmenting this with the introduction of simple state-dependent control procedures provides systems which exhibit more desirable operational features.

  3. Reliability of human-supervised formant-trajectory measurement for forensic voice comparison.

    PubMed

    Zhang, Cuiling; Morrison, Geoffrey Stewart; Ochoa, Felipe; Enzinger, Ewald

    2013-01-01

    Acoustic-phonetic approaches to forensic voice comparison often include human-supervised measurement of vowel formants, but the reliability of such measurements is a matter of concern. This study assesses the within- and between-supervisor variability of three sets of formant-trajectory measurements made by each of four human supervisors. It also assesses the validity and reliability of forensic-voice-comparison systems based on these measurements. Each supervisor's formant-trajectory system was fused with a baseline mel-frequency cepstral-coefficient system, and performance was assessed relative to the baseline system. Substantial improvements in validity were found for all supervisors' systems, but some supervisors' systems were more reliable than others.

  4. DLMS Voice Data Entry.

    DTIC Science & Technology

    1980-06-01

    34 LIST OF ILLUSTRATIONS FIGURE PAGE 1 Block Diagram of DLMS Voice Recognition System .............. S 2 Flowchart of DefaulV...particular are a speech preprocessor and a minicomputer. In the VRS, as shown in the block diagram of Fig. 1, the preprocessor is a TTI model 8040 and...Data General 6026 Magnetic Zo 4 Tape Unit Display L-- - Equipment Cabinet Fig. 1 block Diagram of DIMS Voice Recognition System qS 2. Flexible Disk

  5. Effects of an Automated Telephone Support System on Caregiver Burden and Anxiety: Findings from the REACH for TLC Intervention Study

    ERIC Educational Resources Information Center

    Mahoney, Diane Feeney; Tarlow, Barbara J.; Jones, Richard N.

    2003-01-01

    Purpose: We determine the main outcome effects of a 12-month computer-mediated automated interactive voice response (IVR) intervention designed to assist family caregivers managing persons with disruptive behaviors related to Alzheimer's disease (AD). Design and Methods: We conducted a randomized controlled study of 100 caregivers, 51 in the usual…

  6. The recognition of female voice based on voice registers in singing techniques in real-time using hankel transform method and macdonald function

    NASA Astrophysics Data System (ADS)

    Meiyanti, R.; Subandi, A.; Fuqara, N.; Budiman, M. A.; Siahaan, A. P. U.

    2018-03-01

    A singer doesn’t just recite the lyrics of a song, but also with the use of particular sound techniques to make it more beautiful. In the singing technique, more female have a diverse sound registers than male. There are so many registers of the human voice, but the voice registers used while singing, among others, Chest Voice, Head Voice, Falsetto, and Vocal fry. Research of speech recognition based on the female’s voice registers in singing technique is built using Borland Delphi 7.0. Speech recognition process performed by the input recorded voice samples and also in real time. Voice input will result in weight energy values based on calculations using Hankel Transformation method and Macdonald Functions. The results showed that the accuracy of the system depends on the accuracy of sound engineering that trained and tested, and obtained an average percentage of the successful introduction of the voice registers record reached 48.75 percent, while the average percentage of the successful introduction of the voice registers in real time to reach 57 percent.

  7. Generation of surgical pathology report using a 5,000-word speech recognizer.

    PubMed

    Tischler, A S; Martin, M R

    1989-10-01

    Pressures to decrease both turnaround time and operating costs simultaneously have placed conflicting demands on traditional forms of medical transcription. The new technology of voice recognition extends the promise of enabling the pathologist or other medical professional to dictate a correct report and have it printed and/or transmitted to a database immediately. The usefulness of voice recognition systems depends on several factors, including ease of use, reliability, speed, and accuracy. These in turn depend on the general underlying design of the systems and inclusion in the systems of a specific knowledge base appropriate for each application. Development of a good knowledge base requires close collaboration between a domain expert and a knowledge engineer with expertise in voice recognition. The authors have recently completed a knowledge base for surgical pathology using the Kurzweil VoiceReport 5,000-word system.

  8. Voice problems of group fitness instructors: diagnosis, treatment, perceived and experienced attitudes and expectations of the industry.

    PubMed

    Rumbach, Anna F

    2013-11-01

    To determine the anatomical and physiological nature of voice problems and their treatment in those group fitness instructors (GFIs) who have sought a medical diagnosis; the impact of voice disorders on quality of life and their contribution to activity limitations and participation restrictions; and the perceived attitudes and level of support from the industry at large in response to instructor's voice disorders and need for treatment. Prospective self-completion questionnaire design. Thirty-eight individuals (3 males and 35 females) currently active in the Australian fitness industry who had been diagnosed with a voice disorder completed an online self-completion questionnaire administered via SurveyMonkey. Laryngeal pathology included vocal fold nodules (N = 24), vocal fold cysts (N = 2), vocal fold hemorrhage (N = 1), and recurrent chronic laryngitis (N = 3). Eight individuals reported vocal strain and muscle tension dysphonia without concurrent vocal fold pathology. Treatment methods were variable, with 73.68% (N = 28) receiving voice therapy alone, 7.89% (N = 3) having voice therapy in combination with surgery, and 10.53% (N = 4) having voice therapy in conjunction with medication. Three individuals (7.89%) received no treatment for their voice disorder. During treatment, 82% of the cohort altered their teaching practices. Half of the cohort reported that their voice problems led to social withdrawal, decreased job satisfaction, and emotional distress. Greater than 65% also reported being dissatisfied with the level of industry and coworker support during the period of voice recovery. This study identifies that GFIs are susceptible to a number of voice disorders that impact their social and professional lives, and there is a need for more proactive training and advice on voice care for instructors, as well as those in management positions within the industry to address mixed approaches and opinions regarding the importance of voice care. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  9. Using voice to create hospital progress notes: Description of a mobile application and supporting system integrated with a commercial electronic health record.

    PubMed

    Payne, Thomas H; Alonso, W David; Markiel, J Andrew; Lybarger, Kevin; White, Andrew A

    2018-01-01

    We describe the development and design of a smartphone app-based system to create inpatient progress notes using voice, commercial automatic speech recognition software, with text processing to recognize spoken voice commands and format the note, and integration with a commercial EHR. This new system fits hospital rounding workflow and was used to support a randomized clinical trial testing whether use of voice to create notes improves timeliness of note availability, note quality, and physician satisfaction with the note creation process. The system was used to create 709 notes which were placed in the corresponding patient's EHR record. The median time from pressing the Send button to appearance of the formatted note in the Inbox was 8.8 min. It was generally very reliable, accepted by physician users, and secure. This approach provides an alternative to use of keyboard and templates to create progress notes and may appeal to physicians who prefer voice to typing. Copyright © 2017 Elsevier Inc. All rights reserved.

  10. Telephony-based voice pathology assessment using automated speech analysis.

    PubMed

    Moran, Rosalyn J; Reilly, Richard B; de Chazal, Philip; Lacy, Peter D

    2006-03-01

    A system for remotely detecting vocal fold pathologies using telephone-quality speech is presented. The system uses a linear classifier, processing measurements of pitch perturbation, amplitude perturbation and harmonic-to-noise ratio derived from digitized speech recordings. Voice recordings from the Disordered Voice Database Model 4337 system were used to develop and validate the system. Results show that while a sustained phonation, recorded in a controlled environment, can be classified as normal or pathologic with accuracy of 89.1%, telephone-quality speech can be classified as normal or pathologic with an accuracy of 74.2%, using the same scheme. Amplitude perturbation features prove most robust for telephone-quality speech. The pathologic recordings were then subcategorized into four groups, comprising normal, neuromuscular pathologic, physical pathologic and mixed (neuromuscular with physical) pathologic. A separate classifier was developed for classifying the normal group from each pathologic subcategory. Results show that neuromuscular disorders could be detected remotely with an accuracy of 87%, physical abnormalities with an accuracy of 78% and mixed pathology voice with an accuracy of 61%. This study highlights the real possibility for remote detection and diagnosis of voice pathology.

  11. [Applicability of voice acoustic analysis with vocal loading testto diagnostics of occupational voice diseases].

    PubMed

    Niebudek-Bogusz, Ewa; Sliwińska-Kowalska, Mariola

    2006-01-01

    An assessment of the vocal system, as a part of the medical certification of occupational diseases, should be objective and reliable. Therefore, interest in the method of acoustic voice analysis enabling objective assessment of voice parameters is still growing. The aim of the present study was to evaluate the applicability of acoustic analysis with vocal loading test to the diagnostics of occupational voice disorders. The results of acoustic voice analysis were compared using IRIS software for phoniatrics, before and after a 30-min vocal loading test in 35 female teachers with diagnosed occupational voice disorders (group I) and in 31 female teachers with functional dysphonia (group II). In group I, vocal effort produced significant abnormalities in voice acoustic parameters, compared to group II. These included significantly increased mean fundamental frequency (Fo) value (by 11 Hz) and worsened jitter, shimmer and NHR parameters. Also, the percentage of subjects showing abnormalities in voice acoustic analysis was higher in this group. Conducting voice acoustic analysis before and after the vocal loading test makes it possible to objectively confirm irreversible voice impairments in persons with work-related pathologies of the larynx, which is essential for medical certification of occupational voice diseases.

  12. The interaction of tone with voicing and foot structure: evidence from Kera phonetics and phonology

    NASA Astrophysics Data System (ADS)

    Pearce, Mary Dorothy

    This thesis uses acoustic measurements as a basis for the phonological analysis of the interaction of tone with voicing and foot structure in Kera (a Chadic language). In both tone spreading and vowel harmony, the iambic foot acts as a domain for spreading. Further evidence for the foot comes from measurements of duration, intensity and vowel quality. Kera is unusual in combining a tone system with a partially independent metrical system based on iambs. In words containing more than one foot, the foot is the tone bearing unit (TBU), but in shorter words, the TBU is the syllable. In perception and production experiments, results show that Kera speakers, unlike English and French, use the fundamental frequency as the principle cue to 'Voicing" contrast. Voice onset time (VOT) has only a minor role. Historically, tones probably developed from voicing through a process of tonogenesis, but synchronically, the feature voice is no longer contrastive and VOT is used in an enhancing role. Some linguists have claimed that Kera is a key example for their controversial theory of long-distance voicing spread. But as voice is not part of Kera phonology, this thesis gives counter-evidence to the voice spreading claim. An important finding from the experiments is that the phonological grammars are different between village women, men moving to town and town men. These differences are attributed to French contact. The interaction between Kera tone and voicing and contact with French have produced changes from a 2-way voicing contrast, through a 3-way tonal contrast, to a 2-way voicing contrast plus another contrast with short VOT. These diachronic and synchronic tone/voicing facts are analysed using laryngeal features and Optimality Theory. This thesis provides a body of new data, detailed acoustic measurements, and an analysis incorporating current theoretical issues in phonology, which make it of interest to Africanists and theoreticians alike.

  13. [Signs and symptoms of autonomic dysfunction in dysphonic individuals].

    PubMed

    Park, Kelly; Behlau, Mara

    2011-01-01

    To verify the occurrence of signs and symptoms of autonomic nervous system dysfunction in individuals with behavioral dysphonia, and to compare it with the results obtained by individuals without vocal complaints. Participants were 128 adult individuals with ages between 14 and 74 years, divided into two groups: behavioral dysphonia (61 subjects) and without vocal complaints (67 subjects). It was administered the Protocol of Autonomic Dysfunction, containing 46 questions: 22 related to the autonomic nervous system and had no direct relationship with voice, 16 related to both autonomic nervous system and voice, six non-relevant questions, and two reliability questions. There was a higher occurrence of reported neurovegetative signs in the group with behavioral dysphonia, in questions related to voice, such as frequent throat clearing, frequent swallowing need, fatigability when speaking, and sore throat. In questions not directly related to voice, dysphonic individuals presented greater occurrence of three out of 22 symptoms: gas, tinnitus and aerophagia. Both groups presented similar results in questions non-relevant to the autonomic nervous system. Reliability questions needed reformulation. Individuals with behavioral dysphonia present higher occurrence of neurovegetative signs and symptoms, particularly those with direct relationship with voice, indicating greater lability of the autonomic nervous system in these subjects.

  14. Evaluating iPhone recordings for acoustic voice assessment.

    PubMed

    Lin, Emily; Hornibrook, Jeremy; Ormond, Tika

    2012-01-01

    This study examined the viability of using iPhone recordings for acoustic measurements of voice quality. Acoustic measures were compared between voice signals simultaneously recorded from 11 normal speakers (6 females and 5 males) through an iPhone (model A1303, Apple, USA) and a comparison recording system. Comparisons were also conducted between the pre- and post-operative voices recorded from 10 voice patients (4 females and 6 males) through the iPhone. Participants aged between 27 and 79 years. Measures from iPhone and comparison signals were found to be highly correlated. Findings of the effects of vowel type on the selected measures were consistent between the two recording systems and congruent with previous findings. Analysis of the patient data revealed that a selection of acoustic measures, such as vowel space area and voice perturbation measures, consistently demonstrated a positive change following phonosurgery. The present findings indicated that the iPhone device tested was useful for tracking voice changes for clinical management. Preliminary findings regarding factors such as gender and type of pathology suggest that intra-subject, instead of norm-referenced, comparisons of acoustic measures would be more useful in monitoring the progression of a voice disorder or tracking the treatment effect. Copyright © 2012 S. Karger AG, Basel.

  15. Repetitive transcranial magnetic stimulation of Broca's area affects verbal responses to gesture observation.

    PubMed

    Gentilucci, Maurizio; Bernardis, Paolo; Crisi, Girolamo; Dalla Volta, Riccardo

    2006-07-01

    The aim of the present study was to determine whether Broca's area is involved in translating some aspects of arm gesture representations into mouth articulation gestures. In Experiment 1, we applied low-frequency repetitive transcranial magnetic stimulation over Broca's area and over the symmetrical loci of the right hemisphere of participants responding verbally to communicative spoken words, to gestures, or to the simultaneous presentation of the two signals. We performed also sham stimulation over the left stimulation loci. In Experiment 2, we performed the same stimulations as in Experiment 1 to participants responding with words congruent and incongruent with gestures. After sham stimulation voicing parameters were enhanced when responding to communicative spoken words or to gestures as compared to a control condition of word reading. This effect increased when participants responded to the simultaneous presentation of both communicative signals. In contrast, voicing was interfered when the verbal responses were incongruent with gestures. The left stimulation neither induced enhancement on voicing parameters of words congruent with gestures nor interference on words incongruent with gestures. We interpreted the enhancement of the verbal response to gesturing in terms of intention to interact directly. Consequently, we proposed that Broca's area is involved in the process of translating into speech aspects concerning the social intention coded by the gesture. Moreover, we discussed the results in terms of evolution to support the theory [Corballis, M. C. (2002). From hand to mouth: The origins of language. Princeton, NJ: Princeton University Press] proposing spoken language as evolved from an ancient communication system using arm gestures.

  16. Barriers to disseminating brief CBT for voices from a lived experience and clinician perspective

    PubMed Central

    Hazell, Cassie M.; Strauss, Clara; Cavanagh, Kate

    2017-01-01

    Access to psychological therapies continues to be poor for people experiencing psychosis. To address this problem, researchers are developing brief interventions that address the specific symptoms associated with psychosis, i.e., hearing voices. As part of the development work for a brief Cognitive Behaviour Therapy (CBT) intervention for voices we collected qualitative data from people who hear voices (study 1) and clinicians (study 2) on the potential barriers and facilitators to implementation and engagement. Thematic analysis of the responses from both groups revealed a number of anticipated barriers to implementation and engagement. Both groups believed the presenting problem (voices and psychosis symptoms) may impede engagement. Furthermore clinicians identified a lack of resources to be a barrier to implementation. The only facilitator to engagement was reported by people who hear voices who believed a compassionate, experienced and trustworthy therapist would promote engagement. The results are discussed in relation to how these barriers could be addressed in the context of a brief intervention using CBT techniques. PMID:28575094

  17. Assured Information Flow Capping Architecture.

    DTIC Science & Technology

    1985-05-01

    Air Control System Deployment, ESD-TR-71-371, AD 733 584, Electronic Systems Division, AFSC, Hanscom Air Force Base, MA, November 1971. 3. I. Gitman and...H. Frank, "Economic Analysis of Integrated Voice and Data Networks: A Case Study," Proceedings of the IEEE, November 1978. 4. H. Frank and I. Gitman ... Gitman , "Study Shows Packet Switching Best for Voice Traffic, Too," Data Communications, March 1979. ___ "Economic Analysis of Integrated Voice and

  18. Task-Oriented, Naturally Elicited Speech (TONE) Database for the Force Requirements Expert System, Hawaii (FRESH)

    DTIC Science & Technology

    1988-09-01

    Group Subgroup Command and control; Computational linguistics; expert system voice recognition; man- machine interface; U.S. Government 19 Abstract...simulates the characteristics of FRESH on a smaller scale. This study assisted NOSC in developing a voice-recognition, man- machine interface that could...scale. This study assisted NOSC in developing a voice-recogni- tion, man- machine interface that could be used with TONE and upgraded at a later date

  19. The Johns Hopkins Medical Institutions' Premise Distribution Plan

    PubMed Central

    Barta, Wendy; Buckholtz, Howard; Johnston, Mark; Lenhard, Raymond; Tolchin, Stephen; Vienne, Donald

    1987-01-01

    A Premise Distribution Plan is being developed to address the growing voice and data communications needs at Johns Hopkins Medical Institutions. More specifically, the use of a rapidly expanding Ethernet computer network and a new Integrated Services Digital Network (ISDN) Digital Centrex system must be planned to provide easy, reliable and cost-effective data and voice communications services. Existing Premise Distribution Systems are compared along with voice and data technologies which would use them.

  20. Interface Anywhere: Development of a Voice and Gesture System for Spaceflight Operations

    NASA Technical Reports Server (NTRS)

    Thompson, Shelby; Haddock, Maxwell; Overland, David

    2013-01-01

    The Interface Anywhere Project was funded through Innovation Charge Account (ICA) at NASA JSC in the Fall of 2012. The project was collaboration between human factors and engineering to explore the possibility of designing an interface to control basic habitat operations through gesture and voice control; (a) Current interfaces require the users to be physically near an input device in order to interact with the system; and (b) By using voice and gesture commands, the user is able to interact with the system anywhere they want within the work environment.

  1. Voice Messaging.

    ERIC Educational Resources Information Center

    Davis, Barbara D.; Tisdale, Judy Jones; Krapels, Roberta H.

    2001-01-01

    Surveys corporate use of voice message systems by interviewing employees in four different companies. Finds that all four companies viewed their voicemail systems as a supplement to personal contact (not a replacement) and provided training, but had no formal method to assess customer satisfaction with their system. Suggests business communication…

  2. Impact of voice- and knowledge-enabled clinical reporting--US example.

    PubMed

    Bushko, Renata G; Havlicek, Penny L; Deppert, Edward; Epner, Stephen

    2002-01-01

    This study shows qualitative and quantitative estimates of the national and the clinic level impact of utilizing voice and knowledge enabled clinical reporting systems. Using common sense estimation methodology, we show that the delivery of health care can experience a dramatic improvement in four areas as a result of the broad use of voice and knowledge enabled clinical reporting: (1) Process Quality as measured by cost savings, (2) Organizational Quality as measured by compliance, (3) Clinical Quality as measured by clinical outcomes and (4) Service Quality as measured by patient satisfaction. If only 15 percent of US physicians replaced transcription with modem clinical reporting voice-based methodology, about one half billion dollars could be saved. $6.7 Billion could be saved annually if all medical reporting currently transcribed was handled with voice-and knowledge-enabled dictation and reporting systems.

  3. Response to Reidun Tangen

    ERIC Educational Resources Information Center

    Lewis, Ann

    2008-01-01

    Reidun Tangen begins by reviewing interest in children's "voice" (encompassing the consumer driven, rights based, etc). The main body of her paper examines the philosophical underpinnings of child voice in the research context and, in particular, various interpretations of "the subject" (i.e., the knower) and what it is that is known (i.e., the…

  4. Multi-modal demands of a smartphone used to place calls and enter addresses during highway driving relative to two embedded systems.

    PubMed

    Reimer, Bryan; Mehler, Bruce; Reagan, Ian; Kidd, David; Dobres, Jonathan

    2016-12-01

    There is limited research on trade-offs in demand between manual and voice interfaces of embedded and portable technologies. Mehler et al. identified differences in driving performance, visual engagement and workload between two contrasting embedded vehicle system designs (Chevrolet MyLink and Volvo Sensus). The current study extends this work by comparing these embedded systems with a smartphone (Samsung Galaxy S4). None of the voice interfaces eliminated visual demand. Relative to placing calls manually, both embedded voice interfaces resulted in less eyes-off-road time than the smartphone. Errors were most frequent when calling contacts using the smartphone. The smartphone and MyLink allowed addresses to be entered using compound voice commands resulting in shorter eyes-off-road time compared with the menu-based Sensus but with many more errors. Driving performance and physiological measures indicated increased demand when performing secondary tasks relative to 'just driving', but were not significantly different between the smartphone and embedded systems. Practitioner Summary: The findings show that embedded system and portable device voice interfaces place fewer visual demands on the driver than manual interfaces, but they also underscore how differences in system designs can significantly affect not only the demands placed on drivers, but also the successful completion of tasks.

  5. Treatment outcomes for professional voice users.

    PubMed

    Wingate, Judith M; Brown, William S; Shrivastav, Rahul; Davenport, Paul; Sapienza, Christine M

    2007-07-01

    Professional voice users comprise 25% to 35% of the U.S. working population. Their voice problems may interfere with job performance and impact costs for both employers and employees. The purpose of this study was to examine treatment outcomes of two specific rehabilitation programs for a group of professional voice users. Eighteen professional voice users participated in this study; half had complaints of throat pain or vocal fatigue (Dysphonia Group), and half were found to have benign vocal fold lesions (Lesion Group). One group received 5 weeks of expiratory muscle strength training followed by six sessions of traditional voice therapy. Treatment order was reversed for the second group. The study was designed as a repeated measures study with independent variables of treatment order, laryngeal diagnosis (lesion vs non-lesion), gender, and time. Dependent variables included maximum expiratory pressure (MEP), Voice Handicap Index (VHI) score, Vocal Rating Scale (VRS) score, Voice Effort Scale score, phonetogram measures, subglottal pressures, and acoustic and perceptual measures. Results showed significant improvements in MEP, VHI scores, and VRS scores, subglottal pressure for loud intensity, phonetogram area, and dynamic range. No significant difference was found between laryngeal diagnosis groups. A significant difference was not observed for treatment order. It was concluded that the combined treatment was responsible for the improvements observed. The results indicate that a combined modality treatment may be successful in the remediation of vocal problems for professional voice users.

  6. Effects on vocal range and voice quality of singing voice training: the classically trained female voice.

    PubMed

    Pabon, Peter; Stallinga, Rob; Södersten, Maria; Ternström, Sten

    2014-01-01

    A longitudinal study was performed on the acoustical effects of singing voice training under a given study program, using the voice range profile (VRP). Pretraining and posttraining recordings were made of students who participated in a 3-year bachelor singing study program. A questionnaire that included questions on optimal range, register use, classification, vocal health and hygiene, mixing technique, and training goals was used to rate and categorize self-assessed voice changes. Based on the responses, a subgroup of 10 classically trained female voices was selected, which was homogeneous enough for effects of training to be identified. The VRP perimeter contour was analyzed for effects of voice training. Also, a mapping within the VRP of voice quality, as expressed by the crest factor, was used to indicate the register boundaries and to monitor the acoustical consequences of the newly learned vocal technique of "mixed voice." VRPs were averaged across subjects. Findings were compared with the self-assessed vocal changes. Pre/post comparison of the average VRPs showed, in the midrange, (1) a decrease in the VRP area that was associated with the loud chest voice, (2) a reduction of the crest factor values, and (3) a reduction of maximum sound pressure level values. The students' self-evaluations of the voice changes appeared in some cases to contradict the VRP findings. VRPs of individual voices were seen to change over the course of a singing education. These changes were manifest also in the average group. High-resolution computerized recording, complemented with an acoustic register marker, allows a meaningful assessment of some effects of training, on an individual basis and for groups that comprise singers of a specific genre. It is argued that this kind of investigation is possible only within a focused training program, given by a faculty who has agreed on the goals. Copyright © 2014 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  7. Voice Recognition Software Accuracy with Second Language Speakers of English.

    ERIC Educational Resources Information Center

    Coniam, D.

    1999-01-01

    Explores the potential of the use of voice-recognition technology with second-language speakers of English. Involves the analysis of the output produced by a small group of very competent second-language subjects reading a text into the voice recognition software Dragon Systems "Dragon NaturallySpeaking." (Author/VWL)

  8. WES (Waterways Experiment Station) Communications Plan for Voice and Data

    DTIC Science & Technology

    1989-01-01

    modem on a leased line, and two wideband HDLC 56K connections not used on the Honeywell. 30. Honeywell DPS-8 configuration, as of October 1987, is as...based voice system to support additional asynchronous dial-up modem traffic. In June 1987, Dr. N. Radhakhrishnan of the WES Information Technology...voice system (PBX) and very low-speed data communications by the laboratories using 1,200/2,400-baud asynchronous modems over analog phone lines, and

  9. Iris Young's Five Faces of oppression applied to nursing.

    PubMed

    Dubrosky, Rebekah

    2013-01-01

    To explore Iris Young's Five Faces of Oppression as a framework for understanding oppression of nurses. Nursing's gendered status is implicated in nursing's status as a profession. Young's framework clearly lays out the ways in which nurses are oppressed in today's healthcare system. Understanding the structure of nursing's oppression allows nursing to begin to formulate a thoughtful response to oppression and helps nursing find its voice in the larger world of the healthcare system. © 2013 Wiley Periodicals, Inc.

  10. Infusing Technology into Customer Relationships: Balancing High-Tech and High-Touch

    NASA Astrophysics Data System (ADS)

    Salomann, Harald; Kolbe, Lutz; Brenner, Walter

    In today's business environment, self-service is becoming increasingly important. In order to promote their self-service activities, banks have created online-only products and airlines offer exclusive discounts for passengers booking online. Self-service technologies' practical applications demonstrate this approach's potential. For example, Amtrak introduced an IVR (Interactive Voice Response) system, allowing cost savings of 13m; likewise Royal Mail installed an IVR system leading to a reduction of its customer service costs by 25% (Economist 2004).

  11. 33 CFR 157.136 - Two-way voice communications.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... OIL IN BULK Crude Oil Washing (COW) System on Tank Vessels Design, Equipment, and Installation § 157.136 Two-way voice communications. Each tank vessel having a COW system under § 157.10(e), § 157.10a(a...

  12. Voice control of the space shuttle video system

    NASA Technical Reports Server (NTRS)

    Bejczy, A. K.; Dotson, R. S.; Brown, J. W.; Lewis, J. L.

    1981-01-01

    A pilot voice control system developed at the Jet Propulsion Laboratory (JPL) to test and evaluate the feasibility of controlling the shuttle TV cameras and monitors by voice commands utilizes a commercially available discrete word speech recognizer which can be trained to the individual utterances of each operator. Successful ground tests were conducted using a simulated full-scale space shuttle manipulator. The test configuration involved the berthing, maneuvering and deploying a simulated science payload in the shuttle bay. The handling task typically required 15 to 20 minutes and 60 to 80 commands to 4 TV cameras and 2 TV monitors. The best test runs show 96 to 100 percent voice recognition accuracy.

  13. A new VOX technique for reducing noise in voice communication systems. [voice operated keying

    NASA Technical Reports Server (NTRS)

    Morris, C. F.; Morgan, W. C.; Shack, P. E.

    1974-01-01

    A VOX technique for reducing noise in voice communication systems is described which is based on the separation of voice signals into contiguous frequency-band components with the aid of an adaptive VOX in each band. It is shown that this processing scheme can effectively reduce both wideband and narrowband quasi-periodic noise since the threshold levels readjust themselves to suppress noise that exceeds speech components in each band. Results are reported for tests of the adaptive VOX, and it is noted that improvements can still be made in such areas as the elimination of noise pulses, phoneme reproduction at high-noise levels, and the elimination of distortion introduced by phase delay.

  14. Aircraft L-Band Balloon - Simulated Satellite Experiments Volume I: Experiment Description and Voice and Data Modem Test Results

    DOT National Transportation Integrated Search

    1975-10-01

    This report details the result of an experiment performed by the Transportation Systems Center of the Department of Transportation to evaluate candidate voice and data modulation systems for use in an L-Band Air Traffic Control System. The experiment...

  15. PsyScript: a Macintosh application for scripting experiments.

    PubMed

    Bates, Timothy C; D'Oliveiro, Lawrence

    2003-11-01

    PsyScript is a scriptable application allowing users to describe experiments in Apple's compiled high-level object-oriented AppleScript language, while still supporting millisecond or better within-trial event timing (delays can be in milliseconds or refresh-based, and PsyScript can wait on external I/O, such as eye movement fixations). Because AppleScript is object oriented and system-wide, PsyScript experiments support complex branching, code reuse, and integration with other applications. Included AppleScript-based libraries support file handling and stimulus randomization and sampling, as well as more specialized tasks, such as adaptive testing. Advanced features include support for the BBox serial port button box, as well as a low-cost USB-based digital I/O card for millisecond timing, recording of any number and types of responses within a trial, novel responses, such as graphics tablet drawing, and use of the Macintosh sound facilities to provide an accurate voice key, saving voice responses to disk, scriptable image creation, support for flicker-free animation, and gaze-dependent masking. The application is open source, allowing researchers to enhance the feature set and verify internal functions. Both the application and the source are available for free download at www.maccs.mq.edu.au/-tim/psyscript/.

  16. Dragon Stream Cipher for Secure Blackbox Cockpit Voice Recorder

    NASA Astrophysics Data System (ADS)

    Akmal, Fadira; Michrandi Nasution, Surya; Azmi, Fairuz

    2017-11-01

    Aircraft blackbox is a device used to record all aircraft information, which consists of Flight Data Recorder (FDR) and Cockpit Voice Recorder (CVR). Cockpit Voice Recorder contains conversations in the aircraft during the flight.Investigations on aircraft crashes usually take a long time, because it is difficult to find the aircraft blackbox. Then blackbox should have the ability to send information to other places. Aircraft blackbox must have a data security system, data security is a very important part at the time of information exchange process. The system in this research is to perform the encryption and decryption process on Cockpit Voice Recorder by people who are entitled by using Dragon Stream Cipher algorithm. The tests performed are time of data encryption and decryption, and avalanche effect. Result in this paper show us time encryption and decryption are 0,85 seconds and 1,84 second for 30 seconds Cockpit Voice Recorder data witn an avalanche effect 48,67 %.

  17. Voice Biometrics as a Way to Self-service Password Reset

    NASA Astrophysics Data System (ADS)

    Hohgräfe, Bernd; Jacobi, Sebastian

    Password resets are time consuming. Especially when urgent jobs need to be done, it is cumbersome to inform the user helpdesk, to identify oneself and then to wait for response. It is easy to enter a wrong password multiple times, which leads to the blocking of the application. Voice biometrics is an easy and secure way for individuals to reset their own password. Read more about how you can ease the burden of your user helpdesk and how voice biometric password resets benefit your expense situation without harming your security.

  18. Autophonic Loudness of Singers in Simulated Room Acoustic Environments.

    PubMed

    Yadav, Manuj; Cabrera, Densil

    2017-05-01

    This paper aims to study the effect of room acoustics and phonemes on the perception of loudness of one's own voice (autophonic loudness) for a group of trained singers. For a set of five phonemes, 20 singers vocalized over several autophonic loudness ratios, while maintaining pitch constancy over extreme voice levels, within five simulated rooms. There were statistically significant differences in the slope of the autophonic loudness function (logarithm of autophonic loudness as a function of voice sound pressure level) for the five phonemes, with slopes ranging from 1.3 (/a:/) to 2.0 (/z/). There was no significant variation in the autophonic loudness function slopes with variations in room acoustics. The autophonic room response, which represents a systematic decrease in voice levels with increasing levels of room reflections, was also studied, with some evidence found in support. Overall, the average slope of the autophonic room response for the three corner vowels (/a:/, /i:/, and /u:/) was -1.4 for medium autophonic loudness. The findings relating to the slope of the autophonic loudness function are in agreement with the findings of previous studies where the sensorimotor mechanisms in regulating voice were shown to be more important in the perception of autophonic loudness than hearing of room acoustics. However, the role of room acoustics, in terms of the autophonic room response, is shown to be more complicated, requiring further inquiry. Overall, it is shown that autophonic loudness grows at more than twice the rate of loudness growth for sounds created outside the human body. Crown Copyright © 2017. Published by Elsevier Inc. All rights reserved.

  19. Voice interactive electronic warning systems (VIEWS) - An applied approach to voice technology in the helicopter cockpit

    NASA Technical Reports Server (NTRS)

    Voorhees, J. W.; Bucher, N. M.

    1983-01-01

    The cockpit has been one of the most rapidly changing areas of new aircraft design over the past thirty years. In connection with these developments, a pilot can now be considered a decision maker/system manager as well as a vehicle controller. There is, however, a trend towards an information overload in the cockpit, and information processing problems begin to occur for the rotorcraft pilot. One approach to overcome the arising difficulties is based on the utilization of voice technology to improve the information transfer rate in the cockpit with respect to both input and output. Attention is given to the background of speech technology, the application of speech technology within the cockpit, voice interactive electronic warning system (VIEWS) simulation, and methodology. Information subsystems are considered along with a dynamic simulation study, and data collection.

  20. Marshall Space Flight Center Ground Systems Development and Integration

    NASA Technical Reports Server (NTRS)

    Wade, Gina

    2016-01-01

    Ground Systems Development and Integration performs a variety of tasks in support of the Mission Operations Laboratory (MOL) and other Center and Agency projects. These tasks include various systems engineering processes such as performing system requirements development, system architecture design, integration, verification and validation, software development, and sustaining engineering of mission operations systems that has evolved the Huntsville Operations Support Center (HOSC) into a leader in remote operations for current and future NASA space projects. The group is also responsible for developing and managing telemetry and command configuration and calibration databases. Personnel are responsible for maintaining and enhancing their disciplinary skills in the areas of project management, software engineering, software development, software process improvement, telecommunications, networking, and systems management. Domain expertise in the ground systems area is also maintained and includes detailed proficiency in the areas of real-time telemetry systems, command systems, voice, video, data networks, and mission planning systems.

  1. The Enemy's Gospel: Deconstructing Exclusivity and Inventing Inclusivity through the Power of Story

    ERIC Educational Resources Information Center

    Hilder, Monika B.

    2005-01-01

    The problem of exclusivity figures large in education. How can we educate to deconstruct exclusivity and invent inclusivity? This article asserts that an unexamined veneration for the "objective" academic voice is at least partly responsible for the strong tendency to exclusivity, while suggesting that the subjective voice of storytelling can…

  2. Sparking Passion: Engaging Student Voice through Project-Based Learning in Learning Communities

    ERIC Educational Resources Information Center

    Ball, Christy L.

    2016-01-01

    How do we confront entrenched educational practices in higher education that lead to student demotivation, poor retention, and low persistence? This article argues that project-based learning that situates student voice and capacity at the center of culturally-responsive curriculum has the potential to spark student passion for problem-solving…

  3. View from the Shore: Toward an Indian Voice in 1992.

    ERIC Educational Resources Information Center

    Barreiro, Jose

    1990-01-01

    Reviews plans in Spain and the Americas for observances of the 1992 Columbus Quincentenary. Reflects on Indian responses to these observances and resistance to the notion of America's "discovery." Includes testimonies from Indian voices: N. Scott Momaday, Suzan Shown Harjo, Beverly Singer, Ladonna Harris, Rayna Green, and Tim Coulter.…

  4. An automatic speech recognition system with speaker-independent identification support

    NASA Astrophysics Data System (ADS)

    Caranica, Alexandru; Burileanu, Corneliu

    2015-02-01

    The novelty of this work relies on the application of an open source research software toolkit (CMU Sphinx) to train, build and evaluate a speech recognition system, with speaker-independent support, for voice-controlled hardware applications. Moreover, we propose to use the trained acoustic model to successfully decode offline voice commands on embedded hardware, such as an ARMv6 low-cost SoC, Raspberry PI. This type of single-board computer, mainly used for educational and research activities, can serve as a proof-of-concept software and hardware stack for low cost voice automation systems.

  5. Evaluation of Singing Vocal Health in Yakshagana Singers.

    PubMed

    Gunjawate, Dhanshree R; Aithal, Venkataraja U; Devadas, Usha; Guddattu, Vasudeva

    2017-03-01

    Yakshagana, a popular traditional folk art from Karnataka, India, includes singing and dancing. Yakshagana singer or Bhagavata plays an important role in singing and conducting the performance. The present study aims to assess the singing vocal health using Singing Voice Handicap Index-10 (SVHI-10) in these singers and to compare between those who report voice problem and those who do not. A cross-sectional study was carried out on 26 Bhagavata using demographic questionnaire and SVHI-10 in the Kannada language. Descriptive statistics was used to summarize the data. Independent sample t test was used to compare the responses for demographic variables between the two groups of singers with and without voice problems. The difference in scores of SVHI-10 between the two groups was analyzed using Pearson's chi-square test. Of the Bhagavata, 38% reported to have experienced voice problems, which affected their singing, with higher total SVHI-10 score (31.2 ± 5.7) compared with those who did not report any problems (16.81 ± 9.56). A statistically significant difference between the groups was noted in the emotional domain and total scores. The present study provides preliminary information on the voice handicap reported by Bhagavata. The singers reporting voice problems scored higher on SVHI-10. A healthy singing voice is essential for Yakshagana singers, and voice problems can have a significant impact on their performance and livelihood. Hence, results of the present study indicate the need to understand these singers' voice problems and their impact more comprehensively, and educate them about voice care. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  6. Micro-Based Speech Recognition: Instructional Innovation for Handicapped Learners.

    ERIC Educational Resources Information Center

    Horn, Carin E.; Scott, Brian L.

    A new voice based learning system (VBLS), which allows the handicapped user to interact with a microcomputer by voice commands, is described. Speech or voice recognition is the computerized process of identifying a spoken word or phrase, including those resulting from speech impediments. This new technology is helpful to the severely physically…

  7. Listener Perception of Respiratory-Induced Voice Tremor

    ERIC Educational Resources Information Center

    Farinella, Kimberly A.; Hixon, Thomas J.; Hoit, Jeannette D.; Story, Brad H.; Jones, Patricia A.

    2006-01-01

    Purpose: The purpose of this study was to determine the relation of respiratory oscillation to the perception of voice tremor. Method: Forced oscillation of the respiratory system was used to simulate variations in alveolar pressure such as are characteristic of voice tremor of respiratory origin. Five healthy men served as speakers, and 6…

  8. Speech Motor Development during Acquisition of the Voicing Contrast

    ERIC Educational Resources Information Center

    Grigos, Maria I.; Saxman, John H.; Gordon, Andrew M.

    2005-01-01

    Lip and jaw movements were studied longitudinally in 19-month-old children as they acquired the voicing contrast for /p/ and /b/. A movement tracking system obtained lip and jaw kinematics as participants produced the target utterances /papa/ and /baba/. Laryngeal adjustments were also tracked through acoustically recorded voice onset time (VOT)…

  9. Wireless infrared communications for space and terrestrial applications

    NASA Technical Reports Server (NTRS)

    Crimmins, James W.

    1993-01-01

    Voice and data communications via wireless (and fiberless) optical means has been commonplace for many years. However, continuous advances in optoelectronics and microelectronics have resulted in significant advances in wireless optical communications over the last decade. Wilton has specialized in diffuse infrared voice and data communications since 1979. In 1986, NASA Johnson Space Center invited Wilton to apply its wireless telecommunications and factory floor technology to astronaut voice communications aboard the shuttle. In September, 1988 a special infrared voice communications system flew aboard a 'Discovery' Shuttle mission as a flight experiment. Since then the technology has been further developed, resulting in a general purpose of 2Mbs wireless voice/data LAN which has been tested for a variety of applications including use aboard Spacelab. Funds for Wilton's wireless IR development were provided in part by NASA's Technology Utilization Office and by the NASA Small Business Innovative Research Program. As a consequence, Wilton's commercial product capability has been significantly enhanced to include diffuse infrared wireless LAN's as well as wireless infrared telecommunication systems for voice and data.

  10. Land mobile satellite demonstration system

    NASA Technical Reports Server (NTRS)

    Gooch, Guy M.; Nicholas, David C.

    1988-01-01

    A land mobile satellite demonstration system is described. It ulilizes the INMARSAT MARECS B2 satellite at 26 degrees W. The system provides data transmission using a poll-response protocol with error detection and retransmission at 200 b/s rate. For most tests a 1.8 inch monopole antenna was used, along with a satellite EIRP normally used for four voice channels. A brief summary of the results are given and the overall system consisting of three elements in addition to the satellite (the mobile unit, the base station, and the office terminal and map display) is described. Throughput statistics from one trip are summarized.

  11. More than a feeling: discrete emotions mediate the relationship between relative deprivation and reactions to workplace furloughs.

    PubMed

    Osborne, Danny; Smith, Heather J; Huo, Yuen J

    2012-05-01

    A key insight from investigations of individual relative deprivation (IRD) is that people can experience objective disadvantages differently. In this study, university faculty (N = 953) who reported greater IRD in response to a mandatory furlough (i.e., involuntary pay reductions) were more likely to (a) voice options designed to improve the university (voice), (b) consider leaving their job (exit), and (c) neglect their work responsibilities (neglect), but were (d) less likely to express loyalty to the university (loyalty). Consistent with the emotions literature, (a) anger mediated the relationship between IRD and voice, (b) fear between IRD and exit, (c) sadness between IRD and neglect, and (d) gratitude between IRD and loyalty. IRD was inversely associated with self-reported physical and mental health via these different emotional pathways. These results show how discrete emotions can explain responses to IRD and, in turn, contribute to organizational viability and the health of its members.

  12. Conceptual Sound System Design for Clifford Odets' "GOLDEN BOY"

    NASA Astrophysics Data System (ADS)

    Yang, Yen Chun

    There are two different aspects in the process of sound design, "Arts" and "Science". In my opinion, the sound design should engage both aspects strongly and in interaction with each other. I started the process of designing the sound for GOLDEN BOY by building the city soundscape of New York City in 1937. The scenic design for this piece is designed in the round, putting the audience all around the stage; this gave me a great opportunity to use surround and specialization techniques to transform the space into a different sonic world. My specialization design is composed of two subsystems -- one is the four (4) speakers center cluster diffusing towards the four (4) sections of audience, and the other is the four (4) speakers on the four (4) corners of the theatre. The outside ring provides rich sound source localization and the inside ring provides more support for control of the specialization details. In my design four (4) lavalier microphones are hung under the center iron cage from the four (4) corners of the stage. Each microphone is ten (10) feet above the stage. The signal for each microphone is sent to the two (2) center speakers in the cluster diagonally opposite the microphone. With the appropriate level adjustment of the microphones, the audience will not notice the amplification of the voices; however, through my specialization system, the presence and location of the voices of all actors are preserved for all audiences clearly. With such vocal reinforcements provided by the microphones, I no longer need to worry about overwhelming the dialogue on stage by the underscoring. A successful sound system design should not only provide a functional system, but also take the responsibility of bringing actors' voices to the audience and engaging the audience with the world that we create on stage. By designing a system which reinforces the actors' voices while at the same time providing control over localization of movement of sound effects, I was able not only to make the text present and clear for the audiences, but also to support the storyline strongly through my composed music, environmental soundscapes, and underscoring.

  13. Voice Controlled Wheelchair

    NASA Technical Reports Server (NTRS)

    1977-01-01

    Michael Condon, a quadraplegic from Pasadena, California, demonstrates the NASA-developed voice-controlled wheelchair and its manipulator, which can pick up packages, open doors, turn a TV knob, and perform a variety of other functions. A possible boon to paralyzed and other severely handicapped persons, the chair-manipulator system responds to 35 one-word voice commands, such as "go," "stop," "up," "down," "right," "left," "forward," "backward." The heart of the system is a voice-command analyzer which utilizes a minicomputer. Commands are taught I to the computer by the patient's repeating them a number of times; thereafter the analyzer recognizes commands only in the patient's particular speech pattern. The computer translates commands into electrical signals which activate appropriate motors and cause the desired motion of chair or manipulator. Based on teleoperator and robot technology for space-related programs, the voice-controlled system was developed by Jet Propulsion Laboratory under the joint sponsorship of NASA and the Veterans Administration. The wheelchair-manipulator has been tested at Rancho Los Amigos Hospital, Downey, California, and is being evaluated at the VA Prosthetics Center in New York City.

  14. A new voice rating tool for clinical practice.

    PubMed

    Gould, James; Waugh, Jessica; Carding, Paul; Drinnan, Michael

    2012-07-01

    Perceptual rating of voice quality is a key component in the comprehensive assessment of voice, but there are practical difficulties in making reliable measurements. We have developed the Newcastle Audio Ranking (NeAR) test, a new referential system for the rating of voice parameters. In this article, we present our first results using NeAR. We asked five experts and 11 naive raters to assess 15 male and 15 female voices using the NeAR test. We assessed: validity with respect to the GRBAS scale; interrater reliability; sensitivity to subtle voice differences; and the performance of expert versus naïve raters. There was a uniformly excellent agreement with GRBAS (r=0.87) and interrater agreement (intraclass correlation coefficient=0.86). Considering each GRBAS grade of voice separately, there was still good interrater agreement in NeAR, implying it has good sensitivity to subtle changes. All these results were equally true for expert and naive raters. The NeAR test is a promising new tool in the assessment of voice disorders. Copyright © 2012 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  15. Speech versus manual control of camera functions during a telerobotic task

    NASA Technical Reports Server (NTRS)

    Bierschwale, John M.; Sampaio, Carlos E.; Stuart, Mark A.; Smith, Randy L.

    1989-01-01

    Voice input for control of camera functions was investigated in this study. Objective were to (1) assess the feasibility of a voice-commanded camera control system, and (2) identify factors that differ between voice and manual control of camera functions. Subjects participated in a remote manipulation task that required extensive camera-aided viewing. Each subject was exposed to two conditions, voice and manual input, with a counterbalanced administration order. Voice input was found to be significantly slower than manual input for this task. However, in terms of remote manipulator performance errors and subject preference, there was no difference between modalities. Voice control of continuous camera functions is not recommended. It is believed that the use of voice input for discrete functions, such as multiplexing or camera switching, could aid performance. Hybrid mixes of voice and manual input may provide the best use of both modalities. This report contributes to a better understanding of the issues that affect the design of an efficient human/telerobot interface.

  16. Translational Systems Biology and Voice Pathophysiology

    PubMed Central

    Li, Nicole Y. K.; Abbott, Katherine Verdolini; Rosen, Clark; An, Gary; Hebda, Patricia A.; Vodovotz, Yoram

    2011-01-01

    Objectives/Hypothesis Personalized medicine has been called upon to tailor healthcare to an individual's needs. Evidence-based medicine (EBM) has advocated using randomized clinical trials with large populations to evaluate treatment effects. However, due to large variations across patients, the results are likely not to apply to an individual patient. We suggest that a complementary, systems biology approach using computational modeling may help tackle biological complexity in order to improve ultimate patient care. The purpose of the article is: 1) to review the pros and cons of EBM, and 2) to discuss the alternative systems biology method and present its utility in clinical voice research. Study Design Tutorial Methods Literature review and discussion. Results We propose that translational systems biology can address many of the limitations of EBM pertinent to voice and other health care domains, and thus complement current health research models. In particular, recent work using mathematical modeling suggests that systems biology has the ability to quantify the highly complex biologic processes underlying voice pathophysiology. Recent data support the premise that this approach can be applied specifically in the case of phonotrauma and surgically induced vocal fold trauma, and may have particular power to address personalized medicine. Conclusions We propose that evidence around vocal health and disease be expanded beyond a population-based method to consider more fully issues of complexity and systems interactions, especially in implementing personalized medicine in voice care and beyond. PMID:20025041

  17. Multi-modal demands of a smartphone used to place calls and enter addresses during highway driving relative to two embedded systems

    PubMed Central

    Reimer, Bryan; Mehler, Bruce; Reagan, Ian; Kidd, David; Dobres, Jonathan

    2016-01-01

    Abstract There is limited research on trade-offs in demand between manual and voice interfaces of embedded and portable technologies. Mehler et al. identified differences in driving performance, visual engagement and workload between two contrasting embedded vehicle system designs (Chevrolet MyLink and Volvo Sensus). The current study extends this work by comparing these embedded systems with a smartphone (Samsung Galaxy S4). None of the voice interfaces eliminated visual demand. Relative to placing calls manually, both embedded voice interfaces resulted in less eyes-off-road time than the smartphone. Errors were most frequent when calling contacts using the smartphone. The smartphone and MyLink allowed addresses to be entered using compound voice commands resulting in shorter eyes-off-road time compared with the menu-based Sensus but with many more errors. Driving performance and physiological measures indicated increased demand when performing secondary tasks relative to ‘just driving’, but were not significantly different between the smartphone and embedded systems. Practitioner Summary: The findings show that embedded system and portable device voice interfaces place fewer visual demands on the driver than manual interfaces, but they also underscore how differences in system designs can significantly affect not only the demands placed on drivers, but also the successful completion of tasks. PMID:27110964

  18. Controller/Computer Interface with an Air-Ground Data Link

    DOT National Transportation Integrated Search

    1976-06-01

    This report describes the results of an experiment for evaluating the controller/computer interface in an ARTS III/M&S system modified for use with a simulated digital data link and a voice link utilizing a computer-generated voice system. A modified...

  19. A TDM link with channel coding and digital voice.

    NASA Technical Reports Server (NTRS)

    Jones, M. W.; Tu, K.; Harton, P. L.

    1972-01-01

    The features of a TDM (time-division multiplexed) link model are described. A PCM telemetry sequence was coded for error correction and multiplexed with a digitized voice channel. An all-digital implementation of a variable-slope delta modulation algorithm was used to digitize the voice channel. The results of extensive testing are reported. The measured coding gain and the system performance over a Gaussian channel are compared with theoretical predictions and computer simulations. Word intelligibility scores are reported as a measure of voice channel performance.

  20. Voice Interactive Analysis System Study. Final Report, August 28, 1978 through March 23, 1979.

    ERIC Educational Resources Information Center

    Harry, D. P.; And Others

    The Voice Interactive Analysis System study continued research and development of the LISTEN real-time, minicomputer based connected speech recognition system, within NAVTRAEQUIPCEN'S program of developing automatic speech technology in support of training. An attempt was made to identify the most effective features detected by the TTI-500 model…

  1. Construction site Voice Operated Information System (VOIS) test

    NASA Astrophysics Data System (ADS)

    Lawrence, Debbie J.; Hettchen, William

    1991-01-01

    The Voice Activated Information System (VAIS), developed by USACERL, allows inspectors to verbally log on-site inspection reports on a hand held tape recorder. The tape is later processed by the VAIS, which enters the information into the system's database and produces a written report. The Voice Operated Information System (VOIS), developed by USACERL and Automated Sciences Group, through a ESACERL cooperative research and development agreement (CRDA), is an improved voice recognition system based on the concepts and function of the VAIS. To determine the applicability of the VOIS to Corps of Engineers construction projects, Technology Transfer Test Bad (T3B) funds were provided to the Corps of Engineers National Security Agency (NSA) Area Office (Fort Meade) to procure and implement the VOIS, and to train personnel in its use. This report summarizes the NSA application of the VOIS to quality assurance inspection of radio frequency shielding and to progress payment logs, and concludes that the VOIS is an easily implemented system that can offer improvements when applied to repetitive inspection procedures. Use of VOIS can save time during inspection, improve documentation storage, and provide flexible retrieval of stored information.

  2. Doppler compensation by shifting transmitted object frequency within limits

    NASA Technical Reports Server (NTRS)

    Laughlin, C. R., Jr.; Hollenbaugh, R. C.; Allen, W. K. (Inventor)

    1973-01-01

    A system and method are disclosed for position locating, deriving centralized air traffic control data, and communicating via voice and digital signals between a multiplicity of remote aircraft, including supersonic transports, and a central station. Such communication takes place through a synchronous satellite relay station. Side tone ranging patterns, as well as the digital and voice signals, are modulated on a carrier transmitted from the central station and received on all of the supersonic transports. Each aircraft communicates with the ground stations via a different frequency multiplexed spectrum. Supersonic transport position is derived from a computer at the central station and supplied to a local air traffic controller. Position is determined in response to variable phase information imposed on the side tones at the aircrafts. Common to all of the side tone techniques is Doppler compensation for the supersonic transport velocity.

  3. Mobile Health Devices as Tools for Worldwide Cardiovascular Risk Reduction and Disease Management.

    PubMed

    Piette, John D; List, Justin; Rana, Gurpreet K; Townsend, Whitney; Striplin, Dana; Heisler, Michele

    2015-11-24

    We examined evidence on whether mobile health (mHealth) tools, including interactive voice response calls, short message service, or text messaging, and smartphones, can improve lifestyle behaviors and management related to cardiovascular diseases throughout the world. We conducted a state-of-the-art review and literature synthesis of peer-reviewed and gray literature published since 2004. The review prioritized randomized trials and studies focused on cardiovascular diseases and risk factors, but included other reports when they represented the best available evidence. The search emphasized reports on the potential benefits of mHealth interventions implemented in low- and middle-income countries. Interactive voice response and short message service interventions can improve cardiovascular preventive care in developed countries by addressing risk factors including weight, smoking, and physical activity. Interactive voice response and short message service-based interventions for cardiovascular disease management also have shown benefits with respect to hypertension management, hospital readmissions, and diabetic glycemic control. Multimodal interventions including Web-based communication with clinicians and mHealth-enabled clinical monitoring with feedback also have shown benefits. The evidence regarding the potential benefits of interventions using smartphones and social media is still developing. Studies of mHealth interventions have been conducted in >30 low- and middle-income countries, and evidence to date suggests that programs are feasible and may improve medication adherence and disease outcomes. Emerging evidence suggests that mHealth interventions may improve cardiovascular-related lifestyle behaviors and disease management. Next-generation mHealth programs developed worldwide should be based on evidence-based behavioral theories and incorporate advances in artificial intelligence for adapting systems automatically to patients' unique and changing needs. © 2015 American Heart Association, Inc.

  4. Listening to Young Children's Voices: The Evaluation of a Coding System

    ERIC Educational Resources Information Center

    Tertoolen, Anja; Geldens, Jeannette; van Oers, Bert; Popeijus, Herman

    2015-01-01

    Listening to young children's voices is an issue with increasing relevance for many researchers in the field of early childhood research. At the same time, teachers and researchers are faced with challenges to provide children with possibilities to express their notions, and to find ways of comprehending children's voices. In our research we aim…

  5. The "Parental Voice": How the Infant-Toddler (Zero to Three Years) Education System Should Deal with Parents

    ERIC Educational Resources Information Center

    Plotnik, Ronit

    2013-01-01

    Parenthood is a concrete experience that develops while having a psychological existence in its background. It is heard in two voices simultaneously: the overt, concrete one versus the covert, psychological one. It moves between four intersecting axes, which together create the "Parental Voice" model. Axis 1--Parenthood between fantasy…

  6. The Relationship between Student Voice and Perceptions of Motivation, Attachment, Achievement and School Climate in Davidson and Rutherford Counties

    ERIC Educational Resources Information Center

    Matthews, Sharon Elizabeth

    2010-01-01

    This study investigated the extent to which there were statistically significant relationships between school administrators' systemic implementation of student voice work and student perceptions (i.e. achievement, motivation, attachment and school climate) and PLAN performance. Student voice was defined as students being equal partners in school…

  7. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C [Livermore, CA; Holzrichter, John F [Berkeley, CA; Ng, Lawrence C [Danville, CA

    2006-08-08

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  8. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2004-03-23

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  9. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2006-02-14

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  10. System And Method For Characterizing Voiced Excitations Of Speech And Acoustic Signals, Removing Acoustic Noise From Speech, And Synthesizi

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2006-04-25

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  11. Enhanced Living by Assessing Voice Pathology Using a Co-Occurrence Matrix

    PubMed Central

    Muhammad, Ghulam; Alhamid, Mohammed F.; Hossain, M. Shamim; Almogren, Ahmad S.; Vasilakos, Athanasios V.

    2017-01-01

    A large number of the population around the world suffers from various disabilities. Disabilities affect not only children but also adults of different professions. Smart technology can assist the disabled population and lead to a comfortable life in an enhanced living environment (ELE). In this paper, we propose an effective voice pathology assessment system that works in a smart home framework. The proposed system takes input from various sensors, and processes the acquired voice signals and electroglottography (EGG) signals. Co-occurrence matrices in different directions and neighborhoods from the spectrograms of these signals were obtained. Several features such as energy, entropy, contrast, and homogeneity from these matrices were calculated and fed into a Gaussian mixture model-based classifier. Experiments were performed with a publicly available database, namely, the Saarbrucken voice database. The results demonstrate the feasibility of the proposed system in light of its high accuracy and speed. The proposed system can be extended to assess other disabilities in an ELE. PMID:28146069

  12. Enhanced Living by Assessing Voice Pathology Using a Co-Occurrence Matrix.

    PubMed

    Muhammad, Ghulam; Alhamid, Mohammed F; Hossain, M Shamim; Almogren, Ahmad S; Vasilakos, Athanasios V

    2017-01-29

    A large number of the population around the world suffers from various disabilities. Disabilities affect not only children but also adults of different professions. Smart technology can assist the disabled population and lead to a comfortable life in an enhanced living environment (ELE). In this paper, we propose an effective voice pathology assessment system that works in a smart home framework. The proposed system takes input from various sensors, and processes the acquired voice signals and electroglottography (EGG) signals. Co-occurrence matrices in different directions and neighborhoods from the spectrograms of these signals were obtained. Several features such as energy, entropy, contrast, and homogeneity from these matrices were calculated and fed into a Gaussian mixture model-based classifier. Experiments were performed with a publicly available database, namely, the Saarbrucken voice database. The results demonstrate the feasibility of the proposed system in light of its high accuracy and speed. The proposed system can be extended to assess other disabilities in an ELE.

  13. Tracking and data relay satellite system configuration and tradeoff study. Volume 1: TDRS system summary, part 1

    NASA Technical Reports Server (NTRS)

    1972-01-01

    A Tracking and Data Relay Satellite System (TDRSS) concept for service of low and medium data rate user spacecraft has been defined. The TDRS system uses two geosynchronous dual spin satellites compatible with Delta 2914 to provide command, tracking, and telemetry service between multiple low earth orbiting users and a centrally located ground station. The low data rate user service capability via each TDRS is as follows: (1) forward link at UHF: voice to one user, commands to 20 users (sequential), range and range rate service, and (2) return link at VHF: voice from one user, data from 20 users (simultaneous), range and range rate return signals. The medium data rate user service via each TDRS is as follows: (1) forward link at S band: voice or command and tracking signals to one user, and (2) return link at S band: voice, data and tracking signals from one user "order wire" for high priority service requests (implemented with an earth coverage antenna).

  14. Hearing the voices of women in war through continuous quality improvement.

    PubMed

    Farrell, M; Farrell, J D

    1998-04-01

    During the war in Bosnia-Herzegovina (BIH), the infrastructure of the health care system was destroyed and health officials, health care providers, and women experienced the war from their own perspectives. Using the principles of continuous quality improvement (CQI), these perspectives were honored and served as the basis for an international training program to hear the voices of women in war and the concerns of those who cared for them. With the results of a comprehensive national survey of all partners, the World Health Organization's European Regional Office held a five-day workshop that used CQI methods. The facilitators and 18 nurses, midwives, and physicians from six cantons developed interventions to address the women's needs, as detailed from the women's stories and from the professional perspectives of governmental authorities and health providers. The effort illustrates the ways in which the principles and tools of CQI can be used to capture the needs of the users of the services with those responsible for restructuring the health service and those with overall responsibility for the health of woman throughout the country.

  15. A Comparison of Educator Dispositions to Student Responses on the Kentucky Student Voice Survey

    ERIC Educational Resources Information Center

    Whitis, Julie D.

    2017-01-01

    The primary purpose of this study was to determine if a correlation exists between teacher dispositions, grounded in Perceptual Psychology, and student results on the Kentucky Student Voice Survey (KSVS), a 25-question survey adapted from Cambridge Education's Tripod survey. A correlation was found between teacher dispositions and KSVS question…

  16. Ubiquitous Discussion Forum: Introducing Mobile Phones and Voice Discussion into a Web Discussion Forum

    ERIC Educational Resources Information Center

    Wei, Fu-Hsiang; Chen, Gwo-Dong; Wang, Chin-Yeh; Li, Liang-Yi

    2007-01-01

    Web-based discussion forums enable users to share knowledge in straightforward and popular platforms. However, discussion forums have several problems, such as the lack of immediate delivery and response, the heavily text-based medium, inability to hear expressions of voice and the heuristically created discussion topics which can impede the…

  17. Doing the "Work of Hearing": Girls' Voices in Transnational Educational Development Campaigns

    ERIC Educational Resources Information Center

    Khoja-Moolji, Shenila

    2016-01-01

    There is an increasing focus in transnational campaigns for girls' education and empowerment on highlighting the voices of girls from the global south. These moves are made in response to feminist critiques of said campaigns for not attending to the diverse, multiple and complex lived experiences of girls. This article engages in theorising these…

  18. 47 CFR 90.353 - LMS operations in the 902-928 MHz band.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... band. (b) LMS systems are authorized to transmit status and instructional messages, either voice or non-voice, so long as they are related to the location or monitoring functions of the system. (c) LMS... subparts B and C of this part. (d) Multilateration LMS systems will be authorized on a primary basis within...

  19. 47 CFR 90.353 - LMS operations in the 902-928 MHz band.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... band. (b) LMS systems are authorized to transmit status and instructional messages, either voice or non-voice, so long as they are related to the location or monitoring functions of the system. (c) LMS... subparts B and C of this part. (d) Multilateration LMS systems will be authorized on a primary basis within...

  20. 47 CFR 90.353 - LMS operations in the 902-928 MHz band.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... band. (b) LMS systems are authorized to transmit status and instructional messages, either voice or non-voice, so long as they are related to the location or monitoring functions of the system. (c) LMS... subparts B and C of this part. (d) Multilateration LMS systems will be authorized on a primary basis within...

  1. 47 CFR 90.353 - LMS operations in the 902-928 MHz band.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... band. (b) LMS systems are authorized to transmit status and instructional messages, either voice or non-voice, so long as they are related to the location or monitoring functions of the system. (c) LMS... subparts B and C of this part. (d) Multilateration LMS systems will be authorized on a primary basis within...

  2. 47 CFR 90.353 - LMS operations in the 902-928 MHz band.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... band. (b) LMS systems are authorized to transmit status and instructional messages, either voice or non-voice, so long as they are related to the location or monitoring functions of the system. (c) LMS... subparts B and C of this part. (d) Multilateration LMS systems will be authorized on a primary basis within...

  3. Weather data dissemination to aircraft

    NASA Technical Reports Server (NTRS)

    Mcfarland, Richard H.; Parker, Craig B.

    1990-01-01

    Documentation exists that shows weather to be responsible for approximately 40 percent of all general aviation accidents with fatalities. Weather data products available on the ground are becoming more sophisticated and greater in number. Although many of these data are critical to aircraft safety, they currently must be transmitted verbally to the aircraft. This process is labor intensive and provides a low rate of information transfer. Consequently, the pilot is often forced to make life-critical decisions based on incomplete and outdated information. Automated transmission of weather data from the ground to the aircraft can provide the aircrew with accurate data in near-real time. The current National Airspace System Plan calls for such an uplink capability to be provided by the Mode S Beacon System data link. Although this system has a very advanced data link capability, it will not be capable of providing adequate weather data to all airspace users in its planned configuration. This paper delineates some of the important weather data uplink system requirements, and describes a system which is capable of meeting these requirements. The proposed system utilizes a run-length coding technique for image data compression and a hybrid phase and amplitude modulation technique for the transmission of both voice and weather data on existing aeronautical Very High Frequency (VHF) voice communication channels.

  4. A modulatory effect of male voice pitch on long-term memory in women: evidence of adaptation for mate choice?

    PubMed

    Smith, David S; Jones, Benedict C; Feinberg, David R; Allan, Kevin

    2012-01-01

    From a functionalist perspective, human memory should be attuned to information of adaptive value for one's survival and reproductive fitness. While evidence of sensitivity to survival-related information is growing, specific links between memory and information that could impact upon reproductive fitness have remained elusive. Here, in two experiments, we showed that memory in women is sensitive to male voice pitch, a sexually dimorphic cue important for mate choice because it not only serves as an indicator of genetic quality, but may also signal behavioural traits undesirable in a long-term partner. In Experiment 1, we report that women's visual object memory is significantly enhanced when an object's name is spoken during encoding in a masculinised (i.e., lower-pitch) versus feminised (i.e., higher-pitch) male voice, but that no analogous effect occurs when women listen to other women's voices. Experiment 2 replicated this pattern of results, additionally showing that lowering and raising male voice pitch enhanced and impaired women's memory, respectively, relative to a baseline (i.e., unmanipulated) voice condition. The modulatory effect of sexual dimorphism cues in the male voice may reveal a mate-choice adaptation within women's memory, sculpted by evolution in response to the dilemma posed by the double-edged qualities of male masculinity.

  5. [Research on Control System of an Exoskeleton Upper-limb Rehabilitation Robot].

    PubMed

    Wang, Lulu; Hu, Xin; Hu, Jie; Fang, Youfang; He, Rongrong; Yu, Hongliu

    2016-12-01

    In order to help the patients with upper-limb disfunction go on rehabilitation training,this paper proposed an upper-limb exoskeleton rehabilitation robot with four degrees of freedom(DOF),and realized two control schemes,i.e.,voice control and electromyography control.The hardware and software design of the voice control system was completed based on RSC-4128 chips,which realized the speech recognition technology of a specific person.Besides,this study adapted self-made surface eletromyogram(sEMG)signal extraction electrodes to collect sEMG signals and realized pattern recognition by conducting sEMG signals processing,extracting time domain features and fixed threshold algorithm.In addition,the pulse-width modulation(PWM)algorithm was used to realize the speed adjustment of the system.Voice control and electromyography control experiments were then carried out,and the results showed that the mean recognition rate of the voice control and electromyography control reached 93.1%and 90.9%,respectively.The results proved the feasibility of the control system.This study is expected to lay a theoretical foundation for the further improvement of the control system of the upper-limb rehabilitation robot.

  6. A flight investigation of simulated data-link communications during single-pilot IFR flight. Volume 2: Flight evaluations

    NASA Technical Reports Server (NTRS)

    Parker, J. F., Jr.; Duffy, J. W.

    1982-01-01

    Key problems in single pilot instrument flight operations are in the management of flight data and the processing of cockpit information during conditions of heavy workload. A flight data console was developed to allow simulation of a digital data link to replace the current voice communications stem used in air traffic control. This is a human factors evaluation of a data link communications system to determine how such a system might reduce cockpit workload, improve flight proficiency, and be accepted by general aviation pilots. The need for a voice channel as backup to a digital link is examined. The evaluations cover both airport terminal area operations and full mission instrument flight. Results show that general aviation pilots operate well with a digital data link communications system. The findings indicate that a data link system for pilot/ATC communications, with a backup voice channel, is well accepted by general aviation pilots and is considered to be safer, more efficient, and result in less workload than the current voice system.

  7. Damping effects of magnetic fluids of various saturation magnetization (abstract)

    NASA Astrophysics Data System (ADS)

    Chagnon, Mark

    1990-05-01

    Magnetic fluids have been widely accepted for use in loudspeaker voice coil gaps as viscous dampers and liquid coolants. When applied properly to a voice coil in manufacturing of the loudspeaker, dramatic improvement in frequency response and power handling is observed. Over the past decade, a great deal of study has been given to the effects of damping as a function of fluid viscosity. It is known that the apparent viscosity of a magnetic fluid increases as a function of applied magnetic field, and that the viscosity versus field relationship approximate that of the magnetization versus applied field. At applied magnetic field strength sufficient to cause magnetic saturation of the fluid, no further increase in viscosity with increased magnetic field is observed. In order to provide a better understanding of the second order magnetoviscous damping effects in magnetic fluids used in voice coils and to provide a better loudspeaker design criterion using magnetic fluids, we have studied the effect on damping of several magnetic fluids of the same O field viscosity and of varying saturation magnetization. Magnetic fluids with saturation magnetization ranging from 50 to 450 G and 100 cps viscosity at O applied field were injected into the voice coil gap of a standard midrange loudspeaker. The frequency response over the entire dynamic range of the speaker was measured. The changes in frequency response versus fluid magnetization are reported.

  8. Reproducibility of Automated Voice Range Profiles, a Systematic Literature Review.

    PubMed

    Printz, Trine; Rosenberg, Tine; Godballe, Christian; Dyrvig, Anne-Kirstine; Grøntved, Ågot Møller

    2018-05-01

    Reliable voice range profiles are of great importance when measuring effects and side effects from surgery affecting voice capacity. Automated recording systems are increasingly used, but the reproducibility of results is uncertain. Our objective was to identify and review the existing literature on test-retest accuracy of the automated voice range profile assessment. Systematic review. PubMed, Scopus, Cochrane Library, ComDisDome, Embase, and CINAHL (EBSCO). We conducted a systematic literature search of six databases from 1983 to 2016. The following keywords were used: phonetogram, voice range profile, and acoustic voice analysis. Inclusion criteria were automated recording procedure, healthy voices, and no intervention between test and retest. Test-retest values concerning fundamental frequency and voice intensity were reviewed. Of 483 abstracts, 231 full-text articles were read, resulting in six articles included in the final results. The studies found high reliability, but data are few and heterogeneous. The reviewed articles generally reported high reliability of the voice range profile, and thus clinical usefulness, but uncertainty remains because of low sample sizes and different procedures for selecting, collecting, and analyzing data. More data are needed, and clinical conclusions must be drawn with caution. Copyright © 2018 The Authors. Published by Elsevier Inc. All rights reserved.

  9. The Neighborhood Voice: evaluating a mobile research vehicle for recruiting African Americans to participate in cancer control studies.

    PubMed

    Alcaraz, Kassandra I; Weaver, Nancy L; Andresen, Elena M; Christopher, Kara; Kreuter, Matthew W

    2011-09-01

    The Neighborhood Voice is a vehicle customized for conducting health research in community settings. It brings research studies into neighborhoods affected most by health disparities and reaches groups often underrepresented in research samples. This paper reports on the experience and satisfaction of 599 African American women who participated in research on board the Neighborhood Voice. Using bivariate, psychometric, and logistic regression analyses, we examined responses to a brief post-research survey. Most women (71%) reported that they had never previously participated in research, and two-thirds (68%) rated their Neighborhood Voice experience as excellent. Satisfaction scores were highest among first-time research participants (p < .05). Women's ratings of the Neighborhood Voice on Comfort (OR = 4.9; 95% CI = 3.0, 7.9) and Convenience (OR = 1.8; 95% CI = 1.2, 2.9) significantly predicted having an excellent experience. Mobile research facilities may increase participation among disadvantaged and minority populations. Our brief survey instrument is a model for evaluating such outreach.

  10. Top-Down Modulation of Auditory-Motor Integration during Speech Production: The Role of Working Memory.

    PubMed

    Guo, Zhiqiang; Wu, Xiuqin; Li, Weifeng; Jones, Jeffery A; Yan, Nan; Sheft, Stanley; Liu, Peng; Liu, Hanjun

    2017-10-25

    Although working memory (WM) is considered as an emergent property of the speech perception and production systems, the role of WM in sensorimotor integration during speech processing is largely unknown. We conducted two event-related potential experiments with female and male young adults to investigate the contribution of WM to the neurobehavioural processing of altered auditory feedback during vocal production. A delayed match-to-sample task that required participants to indicate whether the pitch feedback perturbations they heard during vocalizations in test and sample sequences matched, elicited significantly larger vocal compensations, larger N1 responses in the left middle and superior temporal gyrus, and smaller P2 responses in the left middle and superior temporal gyrus, inferior parietal lobule, somatosensory cortex, right inferior frontal gyrus, and insula compared with a control task that did not require memory retention of the sequence of pitch perturbations. On the other hand, participants who underwent extensive auditory WM training produced suppressed vocal compensations that were correlated with improved auditory WM capacity, and enhanced P2 responses in the left middle frontal gyrus, inferior parietal lobule, right inferior frontal gyrus, and insula that were predicted by pretraining auditory WM capacity. These findings indicate that WM can enhance the perception of voice auditory feedback errors while inhibiting compensatory vocal behavior to prevent voice control from being excessively influenced by auditory feedback. This study provides the first evidence that auditory-motor integration for voice control can be modulated by top-down influences arising from WM, rather than modulated exclusively by bottom-up and automatic processes. SIGNIFICANCE STATEMENT One outstanding question that remains unsolved in speech motor control is how the mismatch between predicted and actual voice auditory feedback is detected and corrected. The present study provides two lines of converging evidence, for the first time, that working memory cannot only enhance the perception of vocal feedback errors but also exert inhibitory control over vocal motor behavior. These findings represent a major advance in our understanding of the top-down modulatory mechanisms that support the detection and correction of prediction-feedback mismatches during sensorimotor control of speech production driven by working memory. Rather than being an exclusively bottom-up and automatic process, auditory-motor integration for voice control can be modulated by top-down influences arising from working memory. Copyright © 2017 the authors 0270-6474/17/3710324-11$15.00/0.

  11. Investigation of air transportation technology at Princeton University, 1985

    NASA Technical Reports Server (NTRS)

    Stengel, Robert F.

    1987-01-01

    The program proceeded along five avenues during 1985. Guidance and control strategies for penetration of microbursts and wind shear, application of artificial intelligence in flight control and air traffic control systems, the use of voice recognition in the cockpit, the effects of control saturation on closed-loop stability and response of open-loop unstable aircraft, and computer aided control system design are among the topics briefly considered. Areas of investigation relate to guidance and control of commercial transports as well as general aviation aircraft. Interaction between the flight crew and automatic systems is the subject of principal concern.

  12. Design of Phoneme MIDI Codes Using the MIDI Encoding Tool “Auto-F” and Realizing Voice Synthesizing Functions Based on Musical Sounds

    NASA Astrophysics Data System (ADS)

    Modegi, Toshio

    Using our previously developed audio to MIDI code converter tool “Auto-F”, from given vocal acoustic signals we can create MIDI data, which enable to playback the voice-like signals with a standard MIDI synthesizer. Applying this tool, we are constructing a MIDI database, which consists of previously converted simple harmonic structured MIDI codes from a set of 71 Japanese male and female syllable recorded signals. And we are developing a novel voice synthesizing system based on harmonically synthesizing musical sounds, which can generate MIDI data and playback voice signals with a MIDI synthesizer by giving Japanese plain (kana) texts, referring to the syllable MIDI code database. In this paper, we propose an improved MIDI converter tool, which can produce temporally higher-resolution MIDI codes. Then we propose an algorithm separating a set of 20 consonant and vowel phoneme MIDI codes from 71 syllable MIDI converted codes in order to construct a voice synthesizing system. And, we present the evaluation results of voice synthesizing quality between these separated phoneme MIDI codes and their original syllable MIDI codes by our developed 4-syllable word listening tests.

  13. A report on alterations to the speaking and singing voices of four women following hormonal therapy with virilizing agents.

    PubMed

    Baker, J

    1999-12-01

    Four women aged between 27 and 58 years sought otolaryngological examination due to significant alterations to their voices, the primary concerns being hoarseness in vocal quality, lowering of habitual pitch, difficulty projecting their speaking voices, and loss of control over their singing voices. Otolaryngological examination with a mirror or flexible laryngoscope revealed no apparent abnormality of vocal fold structure or function, and the women were referred for speech pathology with diagnoses of functional dysphonia. Objective acoustic measures using the Kay Visipitch indicated significant lowering of the mean fundamental frequency for each woman, and perceptual analysis of the patients' voices during quiet speaking, projected voice use, and comprehensive singing activities revealed a constellation of features typically noted in the pubescent male. The original diagnoses of a functional dysphonia were queried, prompting further exploration of each woman's medical history, revealing in each case onset of vocal symptoms shortly after commencing treatment for conditions with medications containing virilizing agents (eg, Danocrine (danazol), Deca-Durabolin (nandrolene decanoate), and testosterone). Although some of the vocal symptoms decreased in severity with the influences from 6 months voice therapy and after withdrawal from the drugs, a number of symptoms remained permanent, suggesting each subject had suffered significant alterations in vocal physiology, including muscle tissue changes, muscle coordination dysfunction, and propioceptive dysfunction. This retrospective study is presented in order to illustrate that it was both the projected speaking voice and the singing voice that proved so highly sensitive to the virilization effects. The implications for future prospective research studies and responsible clinical practice are discussed.

  14. Design and Implementation of an Interactive Website for Pediatric Voice Therapy-The Concept of In-Between Care: A Telehealth Model.

    PubMed

    Doarn, Charles R; Zacharias, Stephanie; Keck, Casey Stewart; Tabangin, Meredith; DeAlarcon, Alessandro; Kelchner, Lisa

    2018-06-05

    This article describes the design and implementation of a web-based portal developed to provide supported home practice between weekly voice therapy sessions delivered through telehealth to children with voice disorders. This in-between care consisted of supported home practice that was remotely monitored by speech-language pathologists (SLPs). A web-based voice therapy portal (VTP) was developed as a platform so participants could complete voice therapy home practice by an interdisciplinary team of SLPs (specialized in pediatric voice therapy), telehealth specialists, biomedical informaticians, and interface designers. The VTP was subsequently field tested in a group of children with voice disorders, participating in a larger telehealth study. Building the VTP for supported home practice for pediatric voice therapy was challenging, but successful. Key interactive features of the final site included 11 vocal hygiene questions, traditional voice therapy exercises grouped into levels, audio/visual voice therapy demonstrations, a store-and-retrieval system for voice samples, message/chat function, written guidelines for weekly therapy exercises, and questionnaires for parents to complete after each therapy session. Ten participants (9-14 years of age) diagnosed with a voice disorder were enrolled for eight weekly telehealth voice therapy sessions with follow-up in-between care provided using the VTP. The development and implementation of the VTP as a novel platform for the delivery of voice therapy home practice sessions were effective. We found that a versatile individual, who can work with all project staff (speak the language of both SLPs and information technologists), is essential to the development process. Once the website was established, participants and SLPs effectively utilized the web-based VTP. They found it feasible and useful for needed in-between care and reinforcement of therapeutic exercises.

  15. Modifying the verbal expression of a child with autistic behaviors.

    PubMed

    Hargrave, E; Swisher, L

    1975-06-01

    The Bell and Howell Language Master was used in conjunction with the Monterey Language Program to modify the verbal expression of a nine-year-old boy with autistic behaviors. The goal was to train the child to correctly name up to 10 pictures presented individually. Two training modes were used. For one, the therapist spoke at the time (live voice). For the other, she presented a tape recording of her voice via a Language Master. The results suggested that the child's responses to the Language Master were as good as, if not better than, his responses to the live-voice presentations. In addition, observation indicated that he responded more readily to the Language Master presentations. His spontaneous speech was also noted by independent observers to improve in his classroom and in his home. Possible reasons for the improvement in verbal expression are considered.

  16. Utility and accuracy of perceptual voice and speech distinctions in the diagnosis of Parkinson's disease, PSP and MSA-P.

    PubMed

    Miller, Nick; Nath, Uma; Noble, Emma; Burn, David

    2017-06-01

    To determine if perceptual speech measures distinguish people with Parkinson's disease (PD), multiple system atrophy with predominant parkinsonism (MSA-P) and progressive supranuclear palsy (PSP). Speech-language therapists blind to patient characteristics employed clinical rating scales to evaluate speech/voice in 24 people with clinically diagnosed PD, 17 with PSP and 9 with MSA-P, matched for disease duration (mean 4.9 years, standard deviation 2.2). No consistent intergroup differences appeared on specific speech/voice variables. People with PD were significantly less impaired on overall speech/voice severity. Analyses by severity suggested further investigation around laryngeal, resonance and fluency changes may characterize individual groups. MSA-P and PSP compared with PD were distinguished by severity of speech/voice deterioration, but individual speech/voice parameters failed to consistently differentiate groups.

  17. Depressed mothers' infants are less responsive to faces and voices.

    PubMed

    Field, Tiffany; Diego, Miguel; Hernandez-Reif, Maria

    2009-06-01

    A review of our recent research suggests that infants of depressed mothers appeared to be less responsive to faces and voices as early as the neonatal period. At that time they have shown less orienting to the live face/voice stimulus of the Brazelton scale examiner and to their own and other infants' cry sounds. This lesser responsiveness has been attributed to higher arousal, less attentiveness and less "empathy." Their delayed heart rate decelerations to instrumental and vocal music sounds have also been ascribed to their delayed attention and/or slower processing. Later at 3-6 months they showed less negative responding to their mothers' non-contingent and still-face behavior, suggesting that they were more accustomed to this behavior in their mothers. The less responsive behavior of the depressed mothers was further compounded by their comorbid mood states of anger and anxiety and their difficult interaction styles including withdrawn or intrusive interaction styles and their later authoritarian parenting style. Pregnancy massage was effectively used to reduce prenatal depression and facilitate more optimal neonatal behavior. Interaction coaching was used during the postnatal period to help these dyads with their interactions and ultimately facilitate the infants' development.

  18. Infants of Depressed Mothers Are Less Responsive To Faces and Voices: A Review

    PubMed Central

    Field, Tiffany; Diego, Miguel; Hernandez-Reif, Maria

    2009-01-01

    A review of our recent research suggests that infants of depressed mothers appeared to be less responsive to faces and voices as early as the neonatal period. At that time they have shown less orienting to the live face/voice stimulus of the Brazelton scale examiner and to their own and other infants’ cry sounds. This lesser responsiveness has been attributed to higher arousal, less attentiveness and less “empathy.” Their delayed heart rate decelerations to instrumental and vocal music sounds have also been ascribed to their delayed attention and/or slower processing. Later at 3–6 months they showed less negative responding to their mothers’ non-contingent and still-face behavior, suggesting that they were more accustomed to this behavior in their mothers. The less responsive behavior of the depressed mothers was further compounded by their comorbid mood states of anger and anxiety and their difficult interaction styles including withdrawn or intrusive interaction styles and their later authoritarian parenting style. Pregnancy massage was effectively used to reduce prenatal depression and facilitate more optimal neonatal behavior. Interaction coaching was used during the postnatal period to help these dyads with their interactions and ultimately facilitate the infants’ development PMID:19439359

  19. Satellite voice broadcase system study. Volume 1: Executive summary

    NASA Technical Reports Server (NTRS)

    Horstein, M.

    1985-01-01

    The feasibility of providing Voice of America (VOA) broadcasts by satellite relay was investigated. Satellite voice broadcast systems are described for three different frequency bands: HF, FHV, and L-band. Geostationary satellite configurations are considered for both frequency bands. A system of subsynchronous, circular satellites with an orbit period of 8 hours was developed for the HF band. The VHF broadcasts are provided by a system of Molniya satellites. The satellite designs are limited in size and weight to the capability of the STS/Centaur launch vehicle combination. At L-band, only four geostationary satellites are needed to meet the requirements of the complete broadcast schedule. These satellites are comparable in size and weight to current satellites designed for the direct broadcast of video program material.

  20. Electrophysiological and hemodynamic mismatch responses in rats listening to human speech syllables.

    PubMed

    Mahmoudzadeh, Mahdi; Dehaene-Lambertz, Ghislaine; Wallois, Fabrice

    2017-01-01

    Speech is a complex auditory stimulus which is processed according to several time-scales. Whereas consonant discrimination is required to resolve rapid acoustic events, voice perception relies on slower cues. Humans, right from preterm ages, are particularly efficient to encode temporal cues. To compare the capacities of preterms to those observed in other mammals, we tested anesthetized adult rats by using exactly the same paradigm as that used in preterm neonates. We simultaneously recorded neural (using ECoG) and hemodynamic responses (using fNIRS) to series of human speech syllables and investigated the brain response to a change of consonant (ba vs. ga) and to a change of voice (male vs. female). Both methods revealed concordant results, although ECoG measures were more sensitive than fNIRS. Responses to syllables were bilateral, but with marked right-hemispheric lateralization. Responses to voice changes were observed with both methods, while only ECoG was sensitive to consonant changes. These results suggest that rats more effectively processed the speech envelope than fine temporal cues in contrast with human preterm neonates, in whom the opposite effects were observed. Cross-species comparisons constitute a very valuable tool to define the singularities of the human brain and species-specific bias that may help human infants to learn their native language.

  1. Student Voice and the Perils of Popularity

    ERIC Educational Resources Information Center

    Rudduck, Jean; Fielding, Michael

    2006-01-01

    In this article we suggest that the current popularity of student voice can lead to surface compliance--to a quick response that focuses on "how to do it" rather than a reflective review of "why we might want to do it". We look at the links between student consultation and participation and the legacy of the progressive democratic tradition in our…

  2. "In Charge of the Truffula Seeds": On Children's Literature, Rationality and Children's Voices in Philosophy

    ERIC Educational Resources Information Center

    Johansson, Viktor

    2011-01-01

    In this paper I investigate how philosophy can speak for children and how children can have a voice in philosophy and speak for philosophy. I argue that we should understand children as responsible rational individuals who are involved in their own philosophical inquiries and who can be involved in our own philosophical investigations--not because…

  3. Influence of Self-generated Anchors on the Voice Handicap Index-10 (VHI-10).

    PubMed

    Canals-Fortuny, Elisabet; Vila-Rovira, Josep

    2017-03-01

    The aim of this research is to study whether the presentation of the Voice Handicap Index-10 questionnaire administered at the beginning of the treatment impinged on the results of the responses from the end of the treatment. The questionnaire was administered at the beginning of the treatment to a total of 308 patients. After the treatment, a group of 235 patients answered the questionnaire again without any reference to their responses on the initial administration. The other group of participants, consisting of 73 subjects, completed the questionnaire with the answer sheet of their initial self-assessment in sight. The data obtained show that patients who responded to the anchored answer test show less dispersion and a smaller coefficient of variation (0.90) than those who responded to the nonanchored answer test (coefficient of variation = 1.66). The method of administration of the Voice Handicap Index-10 at the end of a treatment influences the dispersion of the results. We recommend that the patient be anchored to the initial answer sheet while responding to the final self-assessment. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  4. Spasmodic dysphonia follow-up with videolaryngoscopy and voice spectrography during treatment with botulinum toxin.

    PubMed

    Esposito, Marcello; Dubbioso, R; Apisa, P; Allocca, R; Santoro, L; Cesari, U

    2015-09-01

    Spasmodic dysphonia (SD) is a focal dystonia of laryngeal muscles seriously impairing quality of voice. Adductor SD (ADSD) is the most common presentation of this disorder that can be identified by specialized phoniatricians and neurologists firstly on a clinical evaluation and then confirmed by videolaryngoscopy (VL). Botulinum toxin (BTX) injection with electromyographic guidance in muscles around vocal cords is the most effective treatment. Voice Handicap Index (VHI) questionnaire is the main tool to assess dysphonia and response to treatment. Objective of this study is to perform VL and voice spectrography (VS) to confirm the efficacy of BTX injections over time. 13 patients with ADSD were studied with VHI, VL and VS before and after 4 consecutive treatment with onobotulinumtoxin-A. For each treatment vocal improvement was proved by a significant reduction of VHI score and increase of maximum time phonation and harmonic-to-noise ratio while VL showed the absence of spasm in most of patients. No change of the response to BTX was found between injections. This study supports the efficacy of the treatment of SD with BTX with objective measurements and suggests that the efficacy of recurring treatments is stable over time.

  5. How silent is silent reading? Intracerebral evidence for top-down activation of temporal voice areas during reading.

    PubMed

    Perrone-Bertolotti, Marcela; Kujala, Jan; Vidal, Juan R; Hamame, Carlos M; Ossandon, Tomas; Bertrand, Olivier; Minotti, Lorella; Kahane, Philippe; Jerbi, Karim; Lachaux, Jean-Philippe

    2012-12-05

    As you might experience it while reading this sentence, silent reading often involves an imagery speech component: we can hear our own "inner voice" pronouncing words mentally. Recent functional magnetic resonance imaging studies have associated that component with increased metabolic activity in the auditory cortex, including voice-selective areas. It remains to be determined, however, whether this activation arises automatically from early bottom-up visual inputs or whether it depends on late top-down control processes modulated by task demands. To answer this question, we collaborated with four epileptic human patients recorded with intracranial electrodes in the auditory cortex for therapeutic purposes, and measured high-frequency (50-150 Hz) "gamma" activity as a proxy of population level spiking activity. Temporal voice-selective areas (TVAs) were identified with an auditory localizer task and monitored as participants viewed words flashed on screen. We compared neural responses depending on whether words were attended or ignored and found a significant increase of neural activity in response to words, strongly enhanced by attention. In one of the patients, we could record that response at 800 ms in TVAs, but also at 700 ms in the primary auditory cortex and at 300 ms in the ventral occipital temporal cortex. Furthermore, single-trial analysis revealed a considerable jitter between activation peaks in visual and auditory cortices. Altogether, our results demonstrate that the multimodal mental experience of reading is in fact a heterogeneous complex of asynchronous neural responses, and that auditory and visual modalities often process distinct temporal frames of our environment at the same time.

  6. Acoustic Measures of Voice and Physiologic Measures of Autonomic Arousal during Speech as a Function of Cognitive Load.

    PubMed

    MacPherson, Megan K; Abur, Defne; Stepp, Cara E

    2017-07-01

    This study aimed to determine the relationship among cognitive load condition and measures of autonomic arousal and voice production in healthy adults. A prospective study design was conducted. Sixteen healthy young adults (eight men, eight women) produced a sentence containing an embedded Stroop task in each of two cognitive load conditions: congruent and incongruent. In both conditions, participants said the font color of the color words instead of the word text. In the incongruent condition, font color differed from the word text, creating an increase in cognitive load relative to the congruent condition in which font color and word text matched. Three physiologic measures of autonomic arousal (pulse volume amplitude, pulse period, and skin conductance response amplitude) and four acoustic measures of voice (sound pressure level, fundamental frequency, cepstral peak prominence, and low-to-high spectral energy ratio) were analyzed for eight sentence productions in each cognitive load condition per participant. A logistic regression model was constructed to predict the cognitive load condition (congruent or incongruent) using subject as a categorical predictor and the three autonomic measures and four acoustic measures as continuous predictors. It revealed that skin conductance response amplitude, cepstral peak prominence, and low-to-high spectral energy ratio were significantly associated with cognitive load condition. During speech produced under increased cognitive load, healthy young adults show changes in physiologic markers of heightened autonomic arousal and acoustic measures of voice quality. Future work is necessary to examine these measures in older adults and individuals with voice disorders. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  7. Voice over Internet Protocol (VoIP) Technology as a Global Learning Tool: Information Systems Success and Control Belief Perspectives

    ERIC Educational Resources Information Center

    Chen, Charlie C.; Vannoy, Sandra

    2013-01-01

    Voice over Internet Protocol- (VoIP) enabled online learning service providers struggling with high attrition rates and low customer loyalty issues despite VoIP's high degree of system fit for online global learning applications. Effective solutions to this prevalent problem rely on the understanding of system quality, information quality, and…

  8. a Study of Multiplexing Schemes for Voice and Data.

    NASA Astrophysics Data System (ADS)

    Sriram, Kotikalapudi

    Voice traffic variations are characterized by on/off transitions of voice calls, and talkspurt/silence transitions of speakers in conversations. A speaker is known to be in silence for more than half the time during a telephone conversation. In this dissertation, we study some schemes which exploit speaker silences for an efficient utilization of the transmission capacity in integrated voice/data multiplexing and in digital speech interpolation. We study two voice/data multiplexing schemes. In each scheme, any time slots momentarily unutilized by the voice traffic are made available to data. In the first scheme, the multiplexer does not use speech activity detectors (SAD), and hence the voice traffic variations are due to call on/off only. In the second scheme, the multiplexer detects speaker silences using SAD and transmits voice only during talkspurts. The multiplexer with SAD performs digital speech interpolation (DSI) as well as dynamic channel allocation to voice and data. The performance of the two schemes is evaluated using discrete-time modeling and analysis. The data delay performance for the case of English speech is compared with that for the case of Japanese speech. A closed form expression for the mean data message delay is derived for the single-channel single-talker case. In a DSI system, occasional speech losses occur whenever the number of speakers in simultaneous talkspurt exceeds the number of TDM voice channels. In a buffered DSI system, speech loss is further reduced at the cost of delay. We propose a novel fixed-delay buffered DSI scheme. In this scheme, speech fill-in/hangover is not required because there are no variable delays. Hence, all silences that naturally occur in speech are fully utilized. Consequently, a substantial improvement in the DSI performance is made possible. The scheme is modeled and analyzed in discrete -time. Its performance is evaluated in terms of the probability of speech clipping, packet rejection ratio, DSI advantage, and the delay.

  9. 47 CFR 22.1007 - Channels for offshore radiotelephone systems.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... emergency auto alarm and voice transmission pertaining to emergency conditions only. Central Subscriber 488... fixed, surface and/or airborne mobile) as indicated, for emergency auto alarm and voice transmission...

  10. 47 CFR 22.1007 - Channels for offshore radiotelephone systems.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... emergency auto alarm and voice transmission pertaining to emergency conditions only. Central Subscriber 488... fixed, surface and/or airborne mobile) as indicated, for emergency auto alarm and voice transmission...

  11. 47 CFR 22.1007 - Channels for offshore radiotelephone systems.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... emergency auto alarm and voice transmission pertaining to emergency conditions only. Central Subscriber 488... fixed, surface and/or airborne mobile) as indicated, for emergency auto alarm and voice transmission...

  12. Management of the Space Physics Analysis Network (SPAN)

    NASA Technical Reports Server (NTRS)

    Green, James L.; Thomas, Valerie L.; Butler, Todd F.; Peters, David J.; Sisson, Patricia L.

    1990-01-01

    Here, the purpose is to define the operational management structure and to delineate the responsibilities of key Space Physics Analysis Network (SPAN) individuals. The management structure must take into account the large NASA and ESA science research community by giving them a major voice in the operation of the system. Appropriate NASA and ESA interfaces must be provided so that there will be adequate communications facilities available when needed. Responsibilities are delineated for the Advisory Committee, the Steering Committee, the Project Scientist, the Project Manager, the SPAN Security Manager, the Internetwork Manager, the Network Operations Manager, the Remote Site Manager, and others.

  13. Culture modulates the brain response to human expressions of emotion: electrophysiological evidence.

    PubMed

    Liu, Pan; Rigoulot, Simon; Pell, Marc D

    2015-01-01

    To understand how culture modulates on-line neural responses to social information, this study compared how individuals from two distinct cultural groups, English-speaking North Americans and Chinese, process emotional meanings of multi-sensory stimuli as indexed by both behaviour (accuracy) and event-related potential (N400) measures. In an emotional Stroop-like task, participants were presented face-voice pairs expressing congruent or incongruent emotions in conditions where they judged the emotion of one modality while ignoring the other (face or voice focus task). Results indicated that while both groups were sensitive to emotional differences between channels (with lower accuracy and higher N400 amplitudes for incongruent face-voice pairs), there were marked group differences in how intruding facial or vocal cues affected accuracy and N400 amplitudes, with English participants showing greater interference from irrelevant faces than Chinese. Our data illuminate distinct biases in how adults from East Asian versus Western cultures process socio-emotional cues, supplying new evidence that cultural learning modulates not only behaviour, but the neurocognitive response to different features of multi-channel emotion expressions. Copyright © 2014 Elsevier Ltd. All rights reserved.

  14. A system for analysis and classification of voice communications

    NASA Technical Reports Server (NTRS)

    Older, H. J.; Jenney, L. L.; Garland, L.

    1973-01-01

    A method for analysis and classification of verbal communications typically associated with manned space missions or simulations was developed. The study was carried out in two phases. Phase 1 was devoted to identification of crew tasks and activities which require voice communication for accomplishment or reporting. Phase 2 entailed development of a message classification system and a preliminary test of its feasibility. The classification system permits voice communications to be analyzed to three progressively more specific levels of detail and to be described in terms of message content, purpose, and the participants in the information exchange. A coding technique was devised to allow messages to be recorded by an eight-digit number.

  15. Correlational Analysis of Speech Intelligibility Tests and Metrics for Speech Transmission

    DTIC Science & Technology

    2017-12-04

    frequency scale (male voice; normal voice effort) ............................... 4 Fig. 2 Diagram of a speech communication system (Letowski...languages. Consonants contain mostly high frequency (above 1500 Hz) speech energy, but this energy is relatively small in comparison to that of the whole...voices (Letowski et al. 1993). Since the mid- frequency spectral region contains mostly vowel energy while consonants are high frequency sounds, an

  16. Responsive consumerism: empowerment in markets for health plans.

    PubMed

    Elbel, Brian; Schlesinger, Mark

    2009-09-01

    American health policy is increasingly relying on consumerism to improve its performance. This article examines a neglected aspect of medical consumerism: the extent to which consumers respond to problems with their health plans. Using a telephone survey of five thousand consumers conducted in 2002, this article assesses how frequently consumers voice formal grievances or exit from their health plan in response to problems of differing severity. This article also examines the potential impact of this responsiveness on both individuals and the market. In addition, using cross-group comparisons of means and regressions, it looks at how the responses of "empowered" consumers compared with those who are "less empowered." The vast majority of consumers do not formally voice their complaints or exit health plans, even in response to problems with significant consequences. "Empowered" consumers are only minimally more likely to formally voice and no more likely to leave their plan. Moreover, given the greater prevalence of trivial problems, consumers are much more likely to complain or leave their plans because of problems that are not severe. Greater empowerment does not alleviate this. While much of the attention on consumerism has focused on prospective choice, understanding how consumers respond to problems is equally, if not more, important. Relying on consumers' responses as a means to protect individual consumers or influence the market for health plans is unlikely to be successful in its current form.

  17. An Investigation of Multidimensional Voice Program Parameters in Three Different Databases for Voice Pathology Detection and Classification.

    PubMed

    Al-Nasheri, Ahmed; Muhammad, Ghulam; Alsulaiman, Mansour; Ali, Zulfiqar; Mesallam, Tamer A; Farahat, Mohamed; Malki, Khalid H; Bencherif, Mohamed A

    2017-01-01

    Automatic voice-pathology detection and classification systems may help clinicians to detect the existence of any voice pathologies and the type of pathology from which patients suffer in the early stages. The main aim of this paper is to investigate Multidimensional Voice Program (MDVP) parameters to automatically detect and classify the voice pathologies in multiple databases, and then to find out which parameters performed well in these two processes. Samples of the sustained vowel /a/ of normal and pathological voices were extracted from three different databases, which have three voice pathologies in common. The selected databases in this study represent three distinct languages: (1) the Arabic voice pathology database; (2) the Massachusetts Eye and Ear Infirmary database (English database); and (3) the Saarbruecken Voice Database (German database). A computerized speech lab program was used to extract MDVP parameters as features, and an acoustical analysis was performed. The Fisher discrimination ratio was applied to rank the parameters. A t test was performed to highlight any significant differences in the means of the normal and pathological samples. The experimental results demonstrate a clear difference in the performance of the MDVP parameters using these databases. The highly ranked parameters also differed from one database to another. The best accuracies were obtained by using the three highest ranked MDVP parameters arranged according to the Fisher discrimination ratio: these accuracies were 99.68%, 88.21%, and 72.53% for the Saarbruecken Voice Database, the Massachusetts Eye and Ear Infirmary database, and the Arabic voice pathology database, respectively. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  18. A hybrid voice/data modulation for the VHF aeronautical channels

    NASA Technical Reports Server (NTRS)

    Akos, Dennis M.

    1993-01-01

    A method of improving the spectral efficiency of the existing Very High Frequency (VHF) Amplitude Modulation (AM) voice communication channels is proposed. The technique is to phase modulate the existing voice amplitude modulated carrier with digital data. This allows the transmission of digital information over an existing AM voice channel with no change to the existing AM signal format. There is no modification to the existing AM receiver to demodulate the voice signal and an additional receiver module can be added for processing of the digital data. The existing VHF AM transmitter requires only a slight modification for the addition of the digital data signal. The past work in the area is summarized and presented together with an improved system design and the proposed implementation.

  19. An open-label study of sodium oxybate (Xyrem®) in spasmodic dysphonia

    PubMed Central

    Rumbach, Anna F.; Blitzer, Andrew; Frucht, Steven J.; Simonyan, Kristina

    2016-01-01

    Objective Spasmodic dysphonia (SD) is a task-specific laryngeal dystonia that affects speech production. Co-occurring voice tremor (VT) often complicates the diagnosis and clinical management of SD. Treatment of SD and VT is largely limited to botulinum toxin injections into laryngeal musculature; other pharmacological options are not sufficiently developed. Study Design and Methods We conducted an open-label study in 23 SD and 22 SD/VT patients to examine the effects of sodium oxybate (Xyrem®), an oral agent with therapeutic effects similar to those of alcohol in these patients. Blinded randomized analysis of voice and speech samples assessed symptom improvement before and after drug administration. Results Sodium oxybate significantly improved voice symptoms (p = 0.001) primarily by reducing the number of SD-characteristic voice breaks and severity of VT. Sodium oxybate further showed a trend for improving VT symptoms (p = 0.03) in a subset of patients who received successful botulinum toxin injections for the management of their SD symptoms. The drug’s effects were observed approximately 30–40 min after its intake and lasted about 3.5–4 hours. Conclusion Our study demonstrated that sodium oxybate reduced voice symptoms in 82.2% of alcohol-responsive SD patients both with and without co-occurring VT. Our findings suggest that the therapeutic mechanism of sodium oxybate in SD and SD/VT may be linked to that of alcohol and as such sodium oxybate might be beneficial for alcohol-responsive SD and SD/VT patients. PMID:27808415

  20. Modeling and Analysis of Hybrid Cellular/WLAN Systems with Integrated Service-Based Vertical Handoff Schemes

    NASA Astrophysics Data System (ADS)

    Xia, Weiwei; Shen, Lianfeng

    We propose two vertical handoff schemes for cellular network and wireless local area network (WLAN) integration: integrated service-based handoff (ISH) and integrated service-based handoff with queue capabilities (ISHQ). Compared with existing handoff schemes in integrated cellular/WLAN networks, the proposed schemes consider a more comprehensive set of system characteristics such as different features of voice and data services, dynamic information about the admitted calls, user mobility and vertical handoffs in two directions. The code division multiple access (CDMA) cellular network and IEEE 802.11e WLAN are taken into account in the proposed schemes. We model the integrated networks by using multi-dimensional Markov chains and the major performance measures are derived for voice and data services. The important system parameters such as thresholds to prioritize handoff voice calls and queue sizes are optimized. Numerical results demonstrate that the proposed ISHQ scheme can maximize the utilization of overall bandwidth resources with the best quality of service (QoS) provisioning for voice and data services.

  1. Central nervous system control of the laryngeal muscles in humans

    PubMed Central

    Ludlow, Christy L.

    2005-01-01

    Laryngeal muscle control may vary for different functions such as: voice for speech communication, emotional expression during laughter and cry, breathing, swallowing, and cough. This review discusses the control of the human laryngeal muscles for some of these different functions. Sensori-motor aspects of laryngeal control have been studied by eliciting various laryngeal reflexes. The role of audition in learning and monitoring ongoing voice production for speech is well known; while the role of somatosensory feedback is less well understood. Reflexive control systems involving central pattern generators may contribute to swallowing, breathing and cough with greater cortical control during volitional tasks such as voice production for speech. Volitional control is much less well understood for each of these functions and likely involves the integration of cortical and subcortical circuits. The new frontier is the study of the central control of the laryngeal musculature for voice, swallowing and breathing and how volitional and reflexive control systems may interact in humans. PMID:15927543

  2. Adaptive Suppression of Noise in Voice Communications

    NASA Technical Reports Server (NTRS)

    Kozel, David; DeVault, James A.; Birr, Richard B.

    2003-01-01

    A subsystem for the adaptive suppression of noise in a voice communication system effects a high level of reduction of noise that enters the system through microphones. The subsystem includes a digital signal processor (DSP) plus circuitry that implements voice-recognition and spectral- manipulation techniques. The development of the adaptive noise-suppression subsystem was prompted by the following considerations: During processing of the space shuttle at Kennedy Space Center, voice communications among test team members have been significantly impaired in several instances because some test participants have had to communicate from locations with high ambient noise levels. Ear protection for the personnel involved is commercially available and is used in such situations. However, commercially available noise-canceling microphones do not provide sufficient reduction of noise that enters through microphones and thus becomes transmitted on outbound communication links.

  3. Evaluation of Different Speech and Touch Interfaces to In-Vehicle Music Retrieval Systems

    PubMed Central

    Garay-Vega, L.; Pradhan, A. K.; Weinberg, G.; Schmidt-Nielsen, B.; Harsham, B.; Shen, Y.; Divekar, G.; Romoser, M.; Knodler, M.; Fisher, D. L.

    2010-01-01

    In-vehicle music retrieval systems are becoming more and more popular. Previous studies have shown that they pose a real hazard to drivers when the interface is a tactile one which requires multiple entries and a combination of manual control and visual feedback. Voice interfaces exist as an alternative. Such interfaces can require either multiple or single conversational turns. In this study, each of 17 participants between the ages of 18 and 30 years old was asked to use three different music-retrieval systems (one with a multiple entry touch interface, the iPod™, one with a multiple turn voice interface, interface B, and one with a single turn voice interface, interface C) while driving through a virtual world. Measures of secondary task performance, eye behavior, vehicle control, and workload were recorded. When compared with the touch interface, the voice interfaces reduced the total time drivers spent with their eyes off the forward roadway, especially in prolonged glances, as well as both the total number of glances away from the forward roadway and the perceived workload. Furthermore, when compared with driving without a secondary task, both voice interfaces did not significantly impact hazard anticipation, the frequency of long glances away from the forward roadway, or vehicle control. The multiple turn voice interface (B) significantly increased both the time it took drivers to complete the task and the workload. The implications for interface design and safety are discussed. PMID:20380920

  4. Effects of voice style, noise level, and acoustic feedback on objective and subjective voice evaluations

    PubMed Central

    Bottalico, Pasquale; Graetzer, Simone; Hunter, Eric J.

    2015-01-01

    Speakers adjust their vocal effort when communicating in different room acoustic and noise conditions and when instructed to speak at different volumes. The present paper reports on the effects of voice style, noise level, and acoustic feedback on vocal effort, evaluated as sound pressure level, and self-reported vocal fatigue, comfort, and control. Speakers increased their level in the presence of babble and when instructed to talk in a loud style, and lowered it when acoustic feedback was increased and when talking in a soft style. Self-reported responses indicated a preference for the normal style without babble noise. PMID:26723357

  5. Using Ambulatory Voice Monitoring to Investigate Common Voice Disorders: Research Update

    PubMed Central

    Mehta, Daryush D.; Van Stan, Jarrad H.; Zañartu, Matías; Ghassemi, Marzyeh; Guttag, John V.; Espinoza, Víctor M.; Cortés, Juan P.; Cheyne, Harold A.; Hillman, Robert E.

    2015-01-01

    Many common voice disorders are chronic or recurring conditions that are likely to result from inefficient and/or abusive patterns of vocal behavior, referred to as vocal hyperfunction. The clinical management of hyperfunctional voice disorders would be greatly enhanced by the ability to monitor and quantify detrimental vocal behaviors during an individual’s activities of daily life. This paper provides an update on ongoing work that uses a miniature accelerometer on the neck surface below the larynx to collect a large set of ambulatory data on patients with hyperfunctional voice disorders (before and after treatment) and matched-control subjects. Three types of analysis approaches are being employed in an effort to identify the best set of measures for differentiating among hyperfunctional and normal patterns of vocal behavior: (1) ambulatory measures of voice use that include vocal dose and voice quality correlates, (2) aerodynamic measures based on glottal airflow estimates extracted from the accelerometer signal using subject-specific vocal system models, and (3) classification based on machine learning and pattern recognition approaches that have been used successfully in analyzing long-term recordings of other physiological signals. Preliminary results demonstrate the potential for ambulatory voice monitoring to improve the diagnosis and treatment of common hyperfunctional voice disorders. PMID:26528472

  6. Factors associated with voice therapy outcomes in the treatment of presbyphonia.

    PubMed

    Mau, Ted; Jacobson, Barbara H; Garrett, C Gaelyn

    2010-06-01

    Age, vocal fold atrophy, glottic closure pattern, and the burden of medical problems are associated with voice therapy outcomes for presbyphonia. Retrospective. Records of patients seen over a 3-year period at a voice center were screened. Inclusion criteria consisted of age over 55 years, primary complaint of hoarseness, presence of vocal fold atrophy on examination, and absence of laryngeal or neurological pathology. Videostroboscopic examinations on initial presentation were reviewed. Voice therapy outcomes were assessed with the American Speech-Language-Hearing Association National Outcomes Measurement System scale. Statistical analysis was performed with Spearman rank correlation and chi(2) tests. Sixty-seven patients were included in the study. Of the patients, 85% demonstrated improvement with voice therapy. The most common type of glottic closure consisted of a slit gap. Gender or age had no effect on voice therapy outcomes. Larger glottic gaps on initial stroboscopy examination and more pronounced vocal fold atrophy were weakly correlated with less improvement from voice therapy. A weak correlation was also found between the number of chronic medical conditions and poorer outcomes from voice therapy. The degree of clinician-determined improvement in vocal function from voice therapy is independent of patient age but is influenced by the degree of vocal fold atrophy, glottic closure pattern, and the patient's burden of medical problems.

  7. Overview of the Anik C satellites and services

    NASA Astrophysics Data System (ADS)

    Smart, F. H.

    An overview of the important technical characteristics of the Anik C series of Canadian communications satellites is presented. The system was launched as part of the Telesat Communications payload of the Space Shuttle in 1982. Among the services the system will in the near future provide are: a 27 MHz channel bandwidth television service for pay-TV distribution in Canada; two TV channels for hockey broadcasts and a transportable TV system; a heavy-voice route telephone service for five major Canadian cities; and a telephone system for business voice and data communications. Services anticipated for Anik-C satellites later in the decade include a Single Channel Per Carrier (SCPC) voice and data communications system for British Columbia and the Maritime Provinces, and a direct-to-home broadcast service to be sold to television markets in the United States.

  8. Working Conditions and Workplace Barriers to Vocal Health in Primary School Teachers.

    PubMed

    Munier, Caitriona; Farrell, Rory

    2016-01-01

    The purpose of this study was to identify the working conditions and workplace barriers to vocal health in primary school teachers. The relationship between working conditions and voice is analyzed. This is a survey study in 42 randomized schools from a restricted geographical area. An 85-item questionnaire was administered to 550 primary school teachers in 42 schools in Dublin. It was designed to obtain information on demographics, vocal use patterns, vocal health, work organization, working conditions, and teacher's perceptions of the conditions in teaching that might cause a voice problem. The relationship between voice and overstretched work demands, and voice and class size, was examined. A chi-squared test was run to test the null hypothesis that the variables overstretched work demands and voice and class size and voice are independent. Subjects were given the opportunity to give their opinion on their working conditions and on the availability of advice and support within the workplace. A final question sought their opinion on what should be included in a voice care program. A 55% response rate was obtained (n = 304). It was found with 96.52% confidence that the variables overstretched work demands and voice are related. Likewise, it was found that the variables class size and voice are related with 99.97% confidence. There are workplace barriers to vocal health. The working conditions of primary school teachers need to be fully adapted to promote vocal health. Changes by education and health policy makers are needed to achieve this goal. There is a need for future research which focuses on the working conditions of teachers. Copyright © 2016. Published by Elsevier Inc.

  9. Influence of Noise Resulting From the Location and Conditions of Classrooms and Schools in Upper Egypt on Teachers' Voices.

    PubMed

    Phadke, Ketaki Vasant; Abo-Hasseba, Ahmed; Švec, Jan G; Geneid, Ahmed

    2018-05-03

    Teachers are professional voice users, always at high risk of developing voice disorders due to high vocal demand and unfavorable environmental conditions. This study aimed at identifying possible correlations between teachers' voice symptoms and their perception of noise, the location of schools, as well as the location and conditions of their classrooms. One hundred forty teachers (ages 21-56) from schools in Upper Egypt participated in this study. They filled out a questionnaire including questions about the severity and frequency of their voice symptoms, noise perception, and the location and conditions of their schools and classrooms. Questionnaire responses were statistically analyzed to identify possible correlations. There were significant correlations (P < 0.05) between voice symptoms, teachers' noise perception, and noise resulting from the location and conditions of schools and classrooms. Teachers experienced severe dysphonia, neck pain, and increased vocal effort with weekly or daily recurrence. Among the teachers who participated in the study, 24.2% felt they were always in a noisy environment, with 51.4% of the total participants reporting having to raise their voices. The most common sources of noise were from student activities and talking in the teachers' own classrooms (61.4%), noise from adjacent classrooms (52.9%), and road traffic (40.7%). Adverse effect on teachers' voices due to noise from poor school and classroom conditions necessitates solutions for the future improvement of conditions in Egyptian schools. This study may help future studies that focus on developing guidelines for the better planning of Egyptian schools in terms of improved infrastructure and architecture, thus considering the general and vocal health of teachers. Copyright © 2018 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  10. Listening to Children's Voices: Literature and the Arts as Means of Responding to the Effects of War, Terrorism, and Disaster

    ERIC Educational Resources Information Center

    Gangi, Jane M.; Barowsky, Ellis

    2009-01-01

    More and more children are forced to deal with crushing hardships. The responsibilities of adults worldwide to attend to the affected children have never been greater. In this article, the authors first give an overview of the psychological risks for children who experience war, terrorism, and disaster. They then listen to the voices of children…

  11. Training to Use Voice Onset Time as a Cue to Talker Identification Induces a Left-Ear/Right-Hemisphere Processing Advantage

    ERIC Educational Resources Information Center

    Francis, Alexander L.; Driscoll, Courtney

    2006-01-01

    We examined the effect of perceptual training on a well-established hemispheric asymmetry in speech processing. Eighteen listeners were trained to use a within-category difference in voice onset time (VOT) to cue talker identity. Successful learners (n = 8) showed faster response times for stimuli presented only to the left ear than for those…

  12. Noise Robust Speech Recognition Applied to Voice-Driven Wheelchair

    NASA Astrophysics Data System (ADS)

    Sasou, Akira; Kojima, Hiroaki

    2009-12-01

    Conventional voice-driven wheelchairs usually employ headset microphones that are capable of achieving sufficient recognition accuracy, even in the presence of surrounding noise. However, such interfaces require users to wear sensors such as a headset microphone, which can be an impediment, especially for the hand disabled. Conversely, it is also well known that the speech recognition accuracy drastically degrades when the microphone is placed far from the user. In this paper, we develop a noise robust speech recognition system for a voice-driven wheelchair. This system can achieve almost the same recognition accuracy as the headset microphone without wearing sensors. We verified the effectiveness of our system in experiments in different environments, and confirmed that our system can achieve almost the same recognition accuracy as the headset microphone without wearing sensors.

  13. THROUGH-THE EARTH (TTE) SYSTEM AND THE IN-MINE POWER LINE (IMPL) SYSTEM

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zvi H. Meiksin

    Work has progressed on both subsystems: the through-the-earth (TTE) system and the In-Mine Power Line (IMPL) system. After the Lab prototype of the IMPL system was perfected to function satisfactorily, the thrust of the work focused on building a first production prototype that can be installed and tested inside a mine. To obtain multi-channel voice communication through the TTE system, effort has proceeded to compress voice messages and make the format compatible with the power-line interface protocol.

  14. ERP Correlates of Pitch Error Detection in Complex Tone and Voice Auditory Feedback with Missing Fundamental

    PubMed Central

    Behroozmand, Roozbeh; Korzyukov, Oleg; Larson, Charles R.

    2012-01-01

    Previous studies have shown that the pitch of a sound is perceived in the absence of its fundamental frequency (F0), suggesting that a distinct mechanism may resolve pitch based on a pattern that exists between harmonic frequencies. The present study investigated whether such a mechanism is active during voice pitch control. ERPs were recorded in response to +200 cents pitch shifts in the auditory feedback of self-vocalizations and complex tones with and without the F0. The absence of the fundamental induced no difference in ERP latencies. However, a right-hemisphere difference was found in the N1 amplitudes with larger responses to complex tones that included the fundamental compared to when it was missing. The P1 and N1 latencies were shorter in the left hemisphere, and the N1 and P2 amplitudes were larger bilaterally for pitch shifts in voice and complex tones compared with pure tones. These findings suggest hemispheric differences in neural encoding of pitch in sounds with missing fundamental. Data from the present study suggest that the right cortical auditory areas, thought to be specialized for spectral processing, may utilize different mechanisms to resolve pitch in sounds with missing fundamental. The left hemisphere seems to perform faster processing to resolve pitch based on the rate of temporal variations in complex sounds compared with pure tones. These effects indicate that the differential neural processing of pitch in the left and right hemispheres may enable the audio-vocal system to detect temporal and spectral variations in the auditory feedback for vocal pitch control. PMID:22386045

  15. Noise Source Visualization Using a Digital Voice Recorder and Low-Cost Sensors

    PubMed Central

    Cho, Yong Thung

    2018-01-01

    Accurate sound visualization of noise sources is required for optimal noise control. Typically, noise measurement systems require microphones, an analog-digital converter, cables, a data acquisition system, etc., which may not be affordable for potential users. Also, many such systems are not highly portable and may not be convenient for travel. Handheld personal electronic devices such as smartphones and digital voice recorders with relatively lower costs and higher performance have become widely available recently. Even though such devices are highly portable, directly implementing them for noise measurement may lead to erroneous results since such equipment was originally designed for voice recording. In this study, external microphones were connected to a digital voice recorder to conduct measurements and the input received was processed for noise visualization. In this way, a low cost, compact sound visualization system was designed and introduced to visualize two actual noise sources for verification with different characteristics: an enclosed loud speaker and a small air compressor. Reasonable accuracy of noise visualization for these two sources was shown over a relatively wide frequency range. This very affordable and compact sound visualization system can be used for many actual noise visualization applications in addition to educational purposes. PMID:29614038

  16. Multipath/RFI/modulation study for DRSS-RFI problem: Voice coding and intelligibility testing for a satellite-based air traffic control system

    NASA Technical Reports Server (NTRS)

    Birch, J. N.; Getzin, N.

    1971-01-01

    Analog and digital voice coding techniques for application to an L-band satellite-basedair traffic control (ATC) system for over ocean deployment are examined. In addition to performance, the techniques are compared on the basis of cost, size, weight, power consumption, availability, reliability, and multiplexing features. Candidate systems are chosen on the bases of minimum required RF bandwidth and received carrier-to-noise density ratios. A detailed survey of automated and nonautomated intelligibility testing methods and devices is presented and comparisons given. Subjective evaluation of speech system by preference tests is considered. Conclusion and recommendations are developed regarding the selection of the voice system. Likewise, conclusions and recommendations are developed for the appropriate use of intelligibility tests, speech quality measurements, and preference tests with the framework of the proposed ATC system.

  17. Full Duplex, Spread Spectrum Radio System

    NASA Technical Reports Server (NTRS)

    Harvey, Bruce A.

    2000-01-01

    The goal of this project was to support the development of a full duplex, spread spectrum voice communications system. The assembly and testing of a prototype system consisting of a Harris PRISM spread spectrum radio, a TMS320C54x signal processing development board and a Zilog Z80180 microprocessor was underway at the start of this project. The efforts under this project were the development of multiple access schemes, analysis of full duplex voice feedback delays, and the development and analysis of forward error correction (FEC) algorithms. The multiple access analysis involved the selection between code division multiple access (CDMA), frequency division multiple access (FDMA) and time division multiple access (TDMA). Full duplex voice feedback analysis involved the analysis of packet size and delays associated with full loop voice feedback for confirmation of radio system performance. FEC analysis included studies of the performance under the expected burst error scenario with the relatively short packet lengths, and analysis of implementation in the TMS320C54x digital signal processor. When the capabilities and the limitations of the components used were considered, the multiple access scheme chosen was a combination TDMA/FDMA scheme that will provide up to eight users on each of three separate frequencies. Packets to and from each user will consist of 16 samples at a rate of 8,000 samples per second for a total of 2 ms of voice information. The resulting voice feedback delay will therefore be 4 - 6 ms. The most practical FEC algorithm for implementation was a convolutional code with a Viterbi decoder. Interleaving of the bits of each packet will be required to offset the effects of burst errors.

  18. Mobile Communication Devices, Ambient Noise, and Acoustic Voice Measures.

    PubMed

    Maryn, Youri; Ysenbaert, Femke; Zarowski, Andrzej; Vanspauwen, Robby

    2017-03-01

    The ability to move with mobile communication devices (MCDs; ie, smartphones and tablet computers) may induce differences in microphone-to-mouth positioning and use in noise-packed environments, and thus influence reliability of acoustic voice measurements. This study investigated differences in various acoustic voice measures between six recording equipments in backgrounds with low and increasing noise levels. One chain of continuous speech and sustained vowel from 50 subjects with voice disorders (all separated by silence intervals) was radiated and re-recorded in an anechoic chamber with five MCDs and one high-quality recording system. These recordings were acquired in one condition without ambient noise and in four conditions with increased ambient noise. A total of 10 acoustic voice markers were obtained in the program Praat. Differences between MCDs and noise condition were assessed with Friedman repeated-measures test and posthoc Wilcoxon signed-rank tests, both for related samples, after Bonferroni correction. (1) Except median fundamental frequency and seven nonsignificant differences, MCD samples have significantly higher acoustic markers than clinical reference samples in minimal environmental noise. (2) Except median fundamental frequency, jitter local, and jitter rap, all acoustic measures on samples recorded with the reference system experienced significant influence from room noise levels. Fundamental frequency is resistant to recording system, environmental noise, and their combination. All other measures, however, were impacted by both recording system and noise condition, and especially by their combination, often already in the reference/baseline condition without added ambient noise. Caution is therefore warranted regarding implementation of MCDs as clinical recording tools, particularly when applied for treatment outcomes assessments. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  19. Frequency response of the USGS short period telemetered seismic system and its suitability for network studies of local earthquakes

    USGS Publications Warehouse

    Eaton, Jerry P.

    1977-01-01

    The USGS telemetered seismic system was intended primarily to record small to moderate earthquakes (magnitude 0 to 4) at distances of a few km to several hundred km. Its frequency response is such that the recorded background noise at a moderately quite Coast Range site has a relatively flat 'record' spectrum from about 1/3 Hz to about 20 Hz. With the system magnification set so that the background noise is clearly recorded (about 1 mm peak-to-peak) one can anticipate that any seismic signal that exceeds background noise appreciably in this spectral region will be large enough to be seen on the seismogram. This response represents the highest sensitivity and broadest bandwidth that we were able to attain with a 1-Hz seismometer, a simple amplifier VCO employing very low-power integrated circuits, and an 8-channel constant-bandwidth FM subcarrier multiplex system for use with commercial voice-grade phone lines.

  20. Recorded maternal voice for preterm neonates undergoing heel lance.

    PubMed

    Johnston, C Celeste; Filion, Francoise; Nuyt, Anne Monique

    2007-10-01

    To determine if a recording of a mother's voice talking soothingly to her baby is useful in diminishing pain in newborns born between 32 and 36 weeks' gestational age (GA) during routine painful procedures. While maternal skin-to-skin contact has been proven efficacious for diminishing procedural pain in both full-term and preterm neonates, it is often not possible for mothers to be present during a painful procedure. Because auditory development occurs before the third trimester of gestation, it was hypothesized that maternal voice could substitute for maternal presence and be effective in diminishing pain response. Preterm infants between 32 and 36 weeks' GA (n = 20) in the first 10 days of life admitted to 2 urban university-affiliated neonatal intensive care units. Crossover design with random ordering of condition. Following informed consent, an audio recording of the mother talking soothingly to her baby was filtered to simulate the mother's voice traveling through amniotic fluid. A final 10-minute recording of repetition of mothers' talking was recorded with maximum peaks of 70 decibels (dB) and played at levels ranging between 60 and 70 ambient decibels (dbA), selected above recommendations of the American Academy of Pediatrics in order to be heard over high ambient noise in the settings. This was played to her infant by a portable cassette tape player 3 times daily during a 48-hour period after feedings (gavage, bottle, or breast). At the end of 48 hours when blood work was required for clinical purposes, using a crossover design, the infant underwent the heel lancing with or without the recording being played. The order of condition was randomized, and the second condition was within 10 days. The Premature Infant Pain Profile (PIPP) was used as primary outcome. This is a composite measure using heart rate, oxygen saturation, 3 facial actions, behavioral state, and gestational age. This measure has demonstrated reliability and validity indexes. There were no significant differences between groups on the PIPP or any of the individual components of the PIPP except a lower oxygen saturation level in the voice condition following the procedure. The second condition, regardless of whether it was voice or control, had higher heart rate scores and lower oxygen saturation scores even in the prelance baseline and warming phases. Order did not affect PIPP scores or facial actions. Different modalities of maternal presence would appear to be necessary to blunt pain response in infants, and recorded maternal voice alone is not sufficient. The loudness of the recording may have obliterated the infant's ability to discern the mother's voice and may even have been aversive, reflected in decreased oxygen saturation levels in the voice condition. Preterm neonates of 32 to 36 weeks' gestation may become sensitized to painful experiences and show anticipatory physiological response.

  1. Electronic Delivery System: Presentation Features.

    DTIC Science & Technology

    1981-04-01

    THE INFOR’"TiO 1. 0 THE FULNCTIONALITY OF THE PRESENTATIO,’, NOT ITS REPLIC., NATURE IS WHAT COUNTS. S-12 REAL ISM _(CNTD. ) * A SEQUENCE OF...E.G, A MOUSE) IS USED FOR INPUTTINZ RESPONSES, THEY CAN BE VERY EFFICIENT, , S-21 -~i INTERACTION - MECHANISt, S (CONTD.) * TOUCH PANELS -- NATURAL , NO...INTERACTION - MECHANISMS (CONTD, i fm O VOICE INPUT --USED WHERE HANDS OR EYES ARE BUSY (E.G., FOR MAINTENANCE AIDING), -- A NATURAL MEANS OF CO;r UNICATION

  2. Twenty-Channel Voice Response System.

    DTIC Science & Technology

    1981-06-01

    read back as "MIKE" "INDIA" "VICTOR" CHO will be read back as "CHARLIE" " HOTEL " "OSCAR". For some locations, the actual name of the airport will be...PROCESSOR (Cont’d.) C-24 is YES nMk>LENGTH? No a is nWA-nMA+l SKYA (nWA) NO is YES lajt-l? NO 7 is YES SMM (INDA) N UKMUC NO s 2 FIGIM C-3.- SA PR

  3. An Investigation of Flight Deck Data Link in the Terminal Area

    NASA Technical Reports Server (NTRS)

    Martin, Lynne; Lozito, Sandra; Kaneshige, John; Dulchinos, Vicki; Sharma, Shivanjli

    2013-01-01

    The Next Generation Air Transportation System (NextGen) and Europe's Single European Sky ATM Research (SESAR) concepts require an increased use of trajectory-based operations, including extensive strategic air traffic control clearances. The clearances are lengthy and complex, which necessitate data link communications to allow for message permanence and integration into the autoflight systems (i.e., autoload capability). This paper examines the use of flight deck data link communications for strategic and tactical clearance usage in the terminal area. A human-in-the-loop simulation was conducted using a high-fidelity flight deck simulator, with ten commercial flight crews as participants. Data were collected from six flight scenarios in the San Francisco terminal airspace. The variables of interest were ATC message modality (voice v. data link), temporal quality of the message (tactical v. strategic) and message length. Dependent variables were message response times, communication clarifications, communication-related errors, and pilot workload. Response time results were longer in data link compared to voice, a finding that has been consistently revealed in a number of other simulations [1]. In addition, strategic clearances and longer messages resulted in a greater number of clarifications and errors, suggesting an increase in uncertainty of message interpretation for the flight crews when compared to tactical clearances. The implications for strategic and compound clearance usage in NextGen and SESAR are discussed

  4. Improving Higher Education Practice through Student Evaluation Systems: Is the Student Voice Being Heard?

    ERIC Educational Resources Information Center

    Blair, Erik; Valdez Noel, Keisha

    2014-01-01

    Many higher education institutions use student evaluation systems as a way of highlighting course and lecturer strengths and areas for improvement. Globally, the student voice has been increasing in volume, and capitalising on student feedback has been proposed as a means to benefit teacher professional development. This paper examines the student…

  5. 78 FR 63488 - 60-Day Notice of Proposed Information Collection: Grant Drawdown Payment Request/LOCCS/VRS Voice...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-10-24

    ... system. The information collected on the payment voucher will also be used as an internal control measure... LOCCS/VRS voice activated system. The information collected on the form serves also as an internal control measure to ensure the lawful and appropriate disbursement of Federal funds. DATES: Comments Due...

  6. Systems concept for speech technology application in general aviation

    NASA Technical Reports Server (NTRS)

    North, R. A.; Bergeron, H.

    1984-01-01

    The application potential of voice recognition and synthesis circuits for general aviation, single-pilot IFR (SPIFR) situations is examined. The viewpoint of the pilot was central to workload analyses and assessment of the effectiveness of the voice systems. A twin-engine, high performance general aviation aircraft on a cross-country fixed route was employed as the study model. No actual control movements were considered and other possible functions were scored by three IFR-rated instructors. The SPIFR was concluded helpful in alleviating visual and manual workloads during take-off, approach and landing, particularly for data retrieval and entry tasks. Voice synthesis was an aid in alerting a pilot to in-flight problems. It is expected that usable systems will be available within 5 yr.

  7. Temporal signatures of processing voiceness and emotion in sound

    PubMed Central

    Gunter, Thomas C.

    2017-01-01

    Abstract This study explored the temporal course of vocal and emotional sound processing. Participants detected rare repetitions in a stimulus stream comprising neutral and surprised non-verbal exclamations and spectrally rotated control sounds. Spectral rotation preserved some acoustic and emotional properties of the vocal originals. Event-related potentials elicited to unrepeated sounds revealed effects of voiceness and emotion. Relative to non-vocal sounds, vocal sounds elicited a larger centro-parietally distributed N1. This effect was followed by greater positivity to vocal relative to non-vocal sounds beginning with the P2 and extending throughout the recording epoch (N4, late positive potential) with larger amplitudes in female than in male listeners. Emotion effects overlapped with the voiceness effects but were smaller and differed topographically. Voiceness and emotion interacted only for the late positive potential, which was greater for vocal-emotional as compared with all other sounds. Taken together, these results point to a multi-stage process in which voiceness and emotionality are represented independently before being integrated in a manner that biases responses to stimuli with socio-emotional relevance. PMID:28338796

  8. Temporal signatures of processing voiceness and emotion in sound.

    PubMed

    Schirmer, Annett; Gunter, Thomas C

    2017-06-01

    This study explored the temporal course of vocal and emotional sound processing. Participants detected rare repetitions in a stimulus stream comprising neutral and surprised non-verbal exclamations and spectrally rotated control sounds. Spectral rotation preserved some acoustic and emotional properties of the vocal originals. Event-related potentials elicited to unrepeated sounds revealed effects of voiceness and emotion. Relative to non-vocal sounds, vocal sounds elicited a larger centro-parietally distributed N1. This effect was followed by greater positivity to vocal relative to non-vocal sounds beginning with the P2 and extending throughout the recording epoch (N4, late positive potential) with larger amplitudes in female than in male listeners. Emotion effects overlapped with the voiceness effects but were smaller and differed topographically. Voiceness and emotion interacted only for the late positive potential, which was greater for vocal-emotional as compared with all other sounds. Taken together, these results point to a multi-stage process in which voiceness and emotionality are represented independently before being integrated in a manner that biases responses to stimuli with socio-emotional relevance. © The Author (2017). Published by Oxford University Press.

  9. Victoria's Child FIRST and IFS differential response system: progress and issues.

    PubMed

    Lonne, Bob; Brown, Gerry; Wagner, Ingrid; Gillespie, Kerri

    2015-01-01

    Differential response has long been utilized by statutory child protection systems in Australia. This article describes the advent and history of Victoria's differential response system, with a particular focus on the Child FIRST and IFS programme. This program entails a partnership arrangement between the Department of Human Services child protection services and community-based, not-for-profit agencies to provide a diverse range of early intervention and prevention services. The findings of a recent external service system evaluation, a judicial inquiry, and the large-scale Child and Family Services Outcomes Survey of parents/carers perspectives of their service experiences are used to critically examine the effectiveness of this differential response approach. Service-user perspectives of the health and wellbeing of children and families are identified, as well as the recognized implementation issues posing significant challenges for the goal of an integrated partnership system. The need for ongoing reform agendas is highlighted along with the policy, program and structural tensions that exist in differential response systems, which are reliant upon partnerships and shared responsibilities for protecting children and assisting vulnerable families. Suggestions are made for utilizing robust research and evaluation that gives voice to service users and promotes their rights and interests. Copyright © 2014 Elsevier Ltd. All rights reserved.

  10. Voice Over Internet Protocol (VoIP) in a Control Center Environment

    NASA Technical Reports Server (NTRS)

    Pirani, Joseph; Calvelage, Steven

    2010-01-01

    The technology of transmitting voice over data networks has been available for over 10 years. Mass market VoIP services for consumers to make and receive standard telephone calls over broadband Internet networks have grown in the last 5 years. While operational costs are less with VoIP implementations as opposed to time division multiplexing (TDM) based voice switches, is it still advantageous to convert a mission control center s voice system to this newer technology? Marshall Space Flight Center (MSFC) Huntsville Operations Support Center (HOSC) has converted its mission voice services to a commercial product that utilizes VoIP technology. Results from this testing, design, and installation have shown unique considerations that must be addressed before user operations. There are many factors to consider for a control center voice design. Technology advantages and disadvantages were investigated as they refer to cost. There were integration concerns which could lead to complex failure scenarios but simpler integration for the mission infrastructure. MSFC HOSC will benefit from this voice conversion with less product replacement cost, less operations cost and a more integrated mission services environment.

  11. Relationship Between Voice and Motor Disabilities of Parkinson's Disease.

    PubMed

    Majdinasab, Fatemeh; Karkheiran, Siamak; Soltani, Majid; Moradi, Negin; Shahidi, Gholamali

    2016-11-01

    To evaluate voice of Iranian patients with Parkinson's disease (PD) and find any relationship between motor disabilities and acoustic voice parameters as speech motor components. We evaluated 27 Farsi-speaking PD patients and 21 age- and sex-matched healthy persons as control. Motor performance was assessed by the Unified Parkinson's Disease Rating Scale part III and Hoehn and Yahr rating scale in the "on" state. Acoustic voice evaluation, including fundamental frequency (f0), standard deviation of f0, minimum of f0, maximum of f0, shimmer, jitter, and harmonic to noise ratio, was done using the Praat software via /a/ prolongation. No difference was seen between the voice of the patients and the voice of the controls. f0 and its variation had a significant correlation with the duration of the disease, but did not have any relationships with the Unified Parkinson's Disease Rating Scale part III. Only limited relationship was observed between voice and motor disabilities. Tremor is an important main feature of PD that affects motor and phonation systems. Females had an older age at onset, more prolonged disease, and more severe motor disabilities (not statistically significant), but phonation disorders were more frequent in males and showed more relationship with severity of motor disabilities. Voice is affected by PD earlier than many other motor components and is more sensitive to disease progression. Tremor is the most effective part of PD that impacts voice. PD has more effect on voice of male versus female patients. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  12. Raising voices: How sixth graders construct authority and knowledge in argumentative essays

    NASA Astrophysics Data System (ADS)

    Monahan, Mary Elizabeth

    This qualitative classroom-based study documents one teacher-researcher's response to the "voice" debate in composition studies and to the opposing views expressed by Elbow and Bartholomae. The author uses Bakhtin's principle of dialogism, Hymes's theory of communicative competence, as well as Ivanic's discussion of discoursally constructed identities to reconceptualize voice and to redesign writing instruction in her sixth grade classroom. This study shows how students, by redefining and then acting on that voice pedagogy in terms that made sense to them, shaped the author's understanding of what counts as "voiced" writing in non-narrative discourse. Based on a grounded-theory analysis of the twenty-six sixth graders' argumentative essays in science, the author explains voice, not as a property of writers or of texts, but as a process of "knowing together"---a collaborative, but not entirely congenial, exercise of establishing one's authority by talking with, against, and through other voices on the issue. As the results of this study show, the students' "I-Ness" or authorial presence within their texts, was born in a nexus of relationships with "rivals," "allies" and "readers." Given their teacher's injunctions to project confidence and authority in argumentative writing, the students assumed fairly adversarial stances toward these conversational partners throughout their essays. Exaggerating the terms for voiced writing built into the curriculum, the sixth graders produced essays that read more like caricatures than examples of argumentation. Their displays of rhetorical bravado and intellectual aggressiveness, however offsetting to the reader, still enabled these sixth graders to composed voiced essays. This study raises doubts about the value of urging students to sound like their "true selves" or to adopt the formal registers of academe. Students, it seems clear, stand to gain by experimenting with a range of textual identities. The author suggests that voice, as a dialogic process, involves a struggle for meaning---in concert, but also very much in conflict with---other speakers and their intentions.

  13. Making Choices, Taking Chances, Facing Challenges, Managing Change: The Implementation of a Voice/Video/Data Network at the Alliance Library System.

    ERIC Educational Resources Information Center

    Wilford, Valerie J.; Logan, Lee; Bell, Lori; Cloyes, Kay

    The Alliance Library System (ALS) is one of 12 regional library systems in Illinois, providing a full spectrum of support services for 300 member libraries of all types (public, school, academic, and special) located in west central Illinois. This paper describes the process by which ALS implemented a voice/video/data network connecting their four…

  14. Calibration of Clinical Audio Recording and Analysis Systems for Sound Intensity Measurement.

    PubMed

    Maryn, Youri; Zarowski, Andrzej

    2015-11-01

    Sound intensity is an important acoustic feature of voice/speech signals. Yet recordings are performed with different microphone, amplifier, and computer configurations, and it is therefore crucial to calibrate sound intensity measures of clinical audio recording and analysis systems on the basis of output of a sound-level meter. This study was designed to evaluate feasibility, validity, and accuracy of calibration methods, including audiometric speech noise signals and human voice signals under typical speech conditions. Calibration consisted of 3 comparisons between data from 29 measurement microphone-and-computer systems and data from the sound-level meter: signal-specific comparison with audiometric speech noise at 5 levels, signal-specific comparison with natural voice at 3 levels, and cross-signal comparison with natural voice at 3 levels. Intensity measures from recording systems were then linearly converted into calibrated data on the basis of these comparisons, and validity and accuracy of calibrated sound intensity were investigated. Very strong correlations and quasisimilarity were found between calibrated data and sound-level meter data across calibration methods and recording systems. Calibration of clinical sound intensity measures according to this method is feasible, valid, accurate, and representative for a heterogeneous set of microphones and data acquisition systems in real-life circumstances with distinct noise contexts.

  15. Self-contained miniature electronics transceiver provides voice communication in hazardous environment

    NASA Technical Reports Server (NTRS)

    Cribb, H. E.

    1970-01-01

    Two-way wireless voice communications system is automatic, provides freedom of movement, allows for complete awareness of the environment, and does not present any additional hazards such as activation of electromagnetic sensitive devices.

  16. Visual Confirmation of Voice Takeoff Clearance (VICON) Alternative Study

    DOT National Transportation Integrated Search

    1980-05-01

    This report presents the results of a program undertaken to study potential alternatives to the VICON (Visual Confirmation of Voice Takeoff Clearance) System which has undergone operational field tests at Bradley International Airport, Windsor Locks,...

  17. Reliability, stability, and sensitivity to change and impairment in acoustic measures of timing and frequency.

    PubMed

    Vogel, Adam P; Fletcher, Janet; Snyder, Peter J; Fredrickson, Amy; Maruff, Paul

    2011-03-01

    Assessment of the voice for supporting classifications of central nervous system (CNS) impairment requires a different practical, methodological, and statistical framework compared with assessment of the voice to guide decisions about change in the CNS. In experimental terms, an understanding of the stability and sensitivity to change of an assessment protocol is required to guide decisions about CNS change. Five experiments (N = 70) were conducted using a set of commonly used stimuli (eg, sustained vowel, reading, extemporaneous speech) and easily acquired measures (eg, f₀-f₄, percent pause). Stability of these measures was examined through their repeated application in healthy adults over brief and intermediate retest intervals (ie, 30 seconds, 2 hours, and 1 week). Those measures found to be stable were then challenged using an experimental model that reliably changes voice acoustic properties (ie, the Lombard effect). Finally, adults with an established CNS-related motor speech disorder (dysarthria) were compared with healthy controls. Of the 61 acoustic variables studied, 36 showed good stability over all three stability experiments (eg, number of pauses, total speech time, speech rate, f₀-f₄. Of the measures with good stability, a number of frequency measures showed a change in response to increased vocal effort resulting from the Lombard effect challenge. Furthermore, several timing measures significantly separated the control and motor speech impairment groups. Measures with high levels of stability within healthy adults, and those that show sensitivity to change and impairment may prove effective for monitoring changes in CNS functioning. Copyright © 2011 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  18. The specificity of neural responses to music and their relation to voice processing: an fMRI-adaptation study.

    PubMed

    Armony, Jorge L; Aubé, William; Angulo-Perkins, Arafat; Peretz, Isabelle; Concha, Luis

    2015-04-23

    Several studies have identified, using functional magnetic resonance imaging (fMRI), a region within the superior temporal gyrus that preferentially responds to musical stimuli. However, in most cases, significant responses to other complex stimuli, particularly human voice, were also observed. Thus, it remains unknown if the same neurons respond to both stimulus types, albeit with different strengths, or whether the responses observed with fMRI are generated by distinct, overlapping neural populations. To address this question, we conducted an fMRI experiment in which short music excerpts and human vocalizations were presented in a pseudo-random order. Critically, we performed an adaptation-based analysis in which responses to the stimuli were analyzed taking into account the category of the preceding stimulus. Our results confirm the presence of a region in the anterior STG that responds more strongly to music than voice. Moreover, we found a music-specific adaptation effect in this area, consistent with the existence of music-preferred neurons. Lack of differences between musicians and non-musicians argues against an expertise effect. These findings provide further support for neural separability between music and speech within the temporal lobe. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  19. A secure mobile multimedia system to assist emergency response teams.

    PubMed

    Belala, Yacine; Issa, Omneya; Gregoire, Jean-Charles; Wong, James

    2008-08-01

    Long wait times after injury and greater distances to travel between accident scenes and medical facilities contribute to increased, possibly unnecessary deaths. This paper describes a mobile emergency system aimed at reducing mortality by improving the readiness of hospital personnel, therefore allowing for more efficient treatment procedures to be performed when the victim arrives. The system is designed to provide a secure transmission of voice, medical data, and video in real-time over third-generation cellular networks. Test results obtained on a commercial network under real-life conditions demonstrate the ability to effectively transmit medical data over 3G networks, making them a viable option available to healthcare professionals.

  20. Development of the child's voice: premutation, mutation.

    PubMed

    Hacki, T; Heitmüller, S

    1999-10-05

    Voice range profile (VRP) measurement was used to evaluate the vocal capabilities of 180 children aged between 4 and 12 years without voice pathology. There were 10 boys and 10 girls in each age group. Using an automatic VRP measurement system, F0 and SPL dB (lin) were determined and displayed two-dimensionally in real time. The speaking voice, the shouting voice and the singing voice were investigated. The results show that vocal capabilities grow with advancing age, but not continuously. The lowering of the habitual pitch of the speaking voice as well as of the entire speaking pitch range occurs for girls between the ages of 7 and 8, for boys between 8 and 9. A temporary restriction of the minimum vocal intensity of the speaking voice (the ability to speak softly) as well as of the singing voice occurs for girls and for boys at the age of 7-8. A decrease of the maximum speech intensity is found for girls at the age of between 7 and 8, for boys between 8 and 9. A lowering of the pitch as well as of the intensity of the shouting voice occurs for both sexes from the age of 10. In contrast to earlier general opinion we note for girls a stage of premutation (between the age of 7 and 8) with essentially the same changes seen among boys, but 1 year earlier. The beginning of the mutation can be fixed at the age of 10-11 years.

  1. Speech-based Class Attendance

    NASA Astrophysics Data System (ADS)

    Faizel Amri, Umar; Nur Wahidah Nik Hashim, Nik; Hazrin Hany Mohamad Hanif, Noor

    2017-11-01

    In the department of engineering, students are required to fulfil at least 80 percent of class attendance. Conventional method requires student to sign his/her initial on the attendance sheet. However, this method is prone to cheating by having another student signing for their fellow classmate that is absent. We develop our hypothesis according to a verse in the Holy Qur’an (95:4), “We have created men in the best of mould”. Based on the verse, we believe each psychological characteristic of human being is unique and thus, their speech characteristic should be unique. In this paper we present the development of speech biometric-based attendance system. The system requires user’s voice to be installed in the system as trained data and it is saved in the system for registration of the user. The following voice of the user will be the test data in order to verify with the trained data stored in the system. The system uses PSD (Power Spectral Density) and Transition Parameter as the method for feature extraction of the voices. Euclidean and Mahalanobis distances are used in order to verified the user’s voice. For this research, ten subjects of five females and five males were chosen to be tested for the performance of the system. The system performance in term of recognition rate is found to be 60% correct identification of individuals.

  2. A Spot Reminder System for the Visually Impaired Based on a Smartphone Camera

    PubMed Central

    Takizawa, Hotaka; Orita, Kazunori; Aoyagi, Mayumi; Ezaki, Nobuo; Mizuno, Shinji

    2017-01-01

    The present paper proposes a smartphone-camera-based system to assist visually impaired users in recalling their memories related to important locations, called spots, that they visited. The memories are recorded as voice memos, which can be played back when the users return to the spots. Spot-to-spot correspondence is determined by image matching based on the scale invariant feature transform. The main contribution of the proposed system is to allow visually impaired users to associate arbitrary voice memos with arbitrary spots. The users do not need any special devices or systems except smartphones and do not need to remember the spots where the voice memos were recorded. In addition, the proposed system can identify spots in environments that are inaccessible to the global positioning system. The proposed system has been evaluated by two experiments: image matching tests and a user study. The experimental results suggested the effectiveness of the system to help visually impaired individuals, including blind individuals, recall information about regularly-visited spots. PMID:28165403

  3. A Spot Reminder System for the Visually Impaired Based on a Smartphone Camera.

    PubMed

    Takizawa, Hotaka; Orita, Kazunori; Aoyagi, Mayumi; Ezaki, Nobuo; Mizuno, Shinji

    2017-02-04

    The present paper proposes a smartphone-camera-based system to assist visually impaired users in recalling their memories related to important locations, called spots, that they visited. The memories are recorded as voice memos, which can be played back when the users return to the spots. Spot-to-spot correspondence is determined by image matching based on the scale invariant feature transform. The main contribution of the proposed system is to allow visually impaired users to associate arbitrary voice memos with arbitrary spots. The users do not need any special devices or systems except smartphones and do not need to remember the spots where the voice memos were recorded. In addition, the proposed system can identify spots in environments that are inaccessible to the global positioning system. The proposed system has been evaluated by two experiments: image matching tests and a user study. The experimental results suggested the effectiveness of the system to help visually impaired individuals, including blind individuals, recall information about regularly-visited spots.

  4. Taming the fear of voice: Dilemmas in maintaining a high vaccination rate in the Netherlands.

    PubMed

    Geelen, Els; van Vliet, Hans; de Hoogh, Pieter; Horstman, Klasien

    2016-03-01

    In the context of international public debates on vaccination the National Institute for Public Health and the Environment (RIVM), the Dutch public health body responsible for the National Immunization Programme (NIP), fears that the high vaccination rate of children in the Netherlands obscures the many doubts and criticisms parents may have about vaccination. The question arises as to how the robustness of this vaccination rate and the resilience of the NIP can be assessed. To answer this question, we explore the vaccination practices and relationships between professionals and parents using qualitative methods. Drawing on Hirschman's concepts of exit, voice and loyalty, we distinguish between two different approaches to vaccination: one which enforces parental loyalty to the vaccination programme, and one which allows for voice. The analysis shows that due to their lack of voice in the main vaccination setting, parents' considerations are unknown and insight into their loyalty is lacking. We argue that the Dutch vaccination programme is caught between the insecurity of enforced parental loyalty to the NIP and the insecurity of enabling parental voice and negotiating space. We conclude that to increase the resilience of the NIP, experimenting with voice and exit is inevitable. Copyright © 2016 Elsevier Ltd. All rights reserved.

  5. An initial study of voice characteristics of children using two different sound coding strategies in comparison to normal hearing children.

    PubMed

    Coelho, Ana Cristina; Brasolotto, Alcione Ghedini; Bevilacqua, Maria Cecília

    2015-06-01

    To compare some perceptual and acoustic characteristics of the voices of children who use the advanced combination encoder (ACE) or fine structure processing (FSP) speech coding strategies, and to investigate whether these characteristics differ from children with normal hearing. Acoustic analysis of the sustained vowel /a/ was performed using the multi-dimensional voice program (MDVP). Analyses of sequential and spontaneous speech were performed using the real time pitch. Perceptual analyses of these samples were performed using visual-analogic scales of pre-selected parameters. Seventy-six children from three years to five years and 11 months of age participated. Twenty-eight were users of ACE, 23 were users of FSP, and 25 were children with normal hearing. Although both groups with CI presented with some deviated vocal features, the users of ACE presented with voice quality more like children with normal hearing than the users of FSP. Sound processing of ACE appeared to provide better conditions for auditory monitoring of the voice, and consequently, for better control of the voice production. However, these findings need to be further investigated due to the lack of comparative studies published to understand exactly which attributes of sound processing are responsible for differences in performance.

  6. Response to Johannsen, Rump, and Linder's Penetrating a wall of introspection: a critical attrition analysis

    NASA Astrophysics Data System (ADS)

    Rushton, Gregory T.; Criswell, Brett A.

    2013-03-01

    In Penetrating a Wall of Introspection: A Critical Attrition Analysis, Johannsen, Rump, and Linder strive to give a voice to those whose thoughts might otherwise be unheard: students representing the casualties in the conflict surrounding the practices of STEM education, specifically those in the field of physics. Beyond giving those students a voice, they try to filter out and amplify a message that the seven individuals themselves may not have recognized: that the cause of their struggles in their physics programs might not be something innate ( causa materialis in the authors' framework), but might be found outside the individuals ( causa efficiens in the authors' framework). In our response, we attempt to extend the conversation regarding the issues these authors have raised by (1) considering the conditions within the physics community that might exacerbate this situation and (2) exploring from a different perspective the nature of the discourses that will either perpetuate or ameliorate such circumstances. In so doing we seek to provide a more holistic description of the features of the educational system that help to construct the wall of introspection and that might, in turn, be redressed in order to help tear it down—and positively impact attrition rates.

  7. The smartphone and the driver's cognitive workload: A comparison of Apple, Google, and Microsoft's intelligent personal assistants.

    PubMed

    Strayer, David L; Cooper, Joel M; Turrill, Jonna; Coleman, James R; Hopman, Rachel J

    2017-06-01

    The goal of this research was to examine the impact of voice-based interactions using 3 different intelligent personal assistants (Apple's Siri , Google's Google Now for Android phones, and Microsoft's Cortana ) on the cognitive workload of the driver. In 2 experiments using an instrumented vehicle on suburban roadways, we measured the cognitive workload of drivers when they used the voice-based features of each smartphone to place a call, select music, or send text messages. Cognitive workload was derived from primary task performance through video analysis, secondary-task performance using the Detection Response Task (DRT), and subjective mental workload. We found that workload was significantly higher than that measured in the single-task drive. There were also systematic differences between the smartphones: The Google system placed lower cognitive demands on the driver than the Apple and Microsoft systems, which did not differ. Video analysis revealed that the difference in mental workload between the smartphones was associated with the number of system errors, the time to complete an action, and the complexity and intuitiveness of the devices. Finally, surprisingly high levels of cognitive workload were observed when drivers were interacting with the devices: "on-task" workload measures did not systematically differ from that associated with a mentally demanding Operation Span (OSPAN) task. The analysis also found residual costs associated using each of the smartphones that took a significant time to dissipate. The data suggest that caution is warranted in the use of smartphone voice-based technology in the vehicle because of the high levels of cognitive workload associated with these interactions. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  8. A flight investigation of simulated data link communications during single-pilot IFR flight

    NASA Technical Reports Server (NTRS)

    Parker, J. F.; Duffy, J. W.; Christensen, D. G.

    1983-01-01

    A Flight Data Console (FDC) was developed to allow simulation of a digital communications link to replace the current voice communication system used in air traffic control (ATC). The voice system requires manipulation of radio equipment, read-back of clearances, and mental storage of critical information items, all contributing to high workload, particularly during single-pilot operations. This was an inflight study to determine how a digital communications system might reduce cockpit workload, improve flight proficiency, and be accepted by general aviation pilots. Results show that instrument flight, including approach and landing, can be accomplished quite effectively using a digital data link system for ATC communications. All pilots expressed a need for a back-up voice channel. When included, this channel was used sparingly and principally to confirm any item of information about which there might be uncertainty.

  9. Modulation of voice related to tremor and vibrato

    NASA Astrophysics Data System (ADS)

    Lester, Rosemary Anne

    Modulation of voice is a result of physiologic oscillation within one or more components of the vocal system including the breathing apparatus (i.e., pressure supply), the larynx (i.e. sound source), and the vocal tract (i.e., sound filter). These oscillations may be caused by pathological tremor associated with neurological disorders like essential tremor or by volitional production of vibrato in singers. Because the acoustical characteristics of voice modulation specific to each component of the vocal system and the effect of these characteristics on perception are not well-understood, it is difficult to assess individuals with vocal tremor and to determine the most effective interventions for reducing the perceptual severity of the disorder. The purpose of the present studies was to determine how the acoustical characteristics associated with laryngeal-based vocal tremor affect the perception of the magnitude of voice modulation, and to determine if adjustments could be made to the voice source and vocal tract filter to alter the acoustic output and reduce the perception of modulation. This research was carried out using both a computational model of speech production and trained singers producing vibrato to simulate laryngeal-based vocal tremor with different voice source characteristics (i.e., vocal fold length and degree of vocal fold adduction) and different vocal tract filter characteristics (i.e., vowel shapes). It was expected that, by making adjustments to the voice source and vocal tract filter that reduce the amplitude of the higher harmonics, the perception of magnitude of voice modulation would be reduced. The results of this study revealed that listeners' perception of the magnitude of modulation of voice was affected by the degree of vocal fold adduction and the vocal tract shape with the computational model, but only by the vocal quality (corresponding to the degree of vocal fold adduction) with the female singer. Based on regression analyses, listeners' judgments were predicted by modulation information in both low and high frequency bands. The findings from these studies indicate that production of a breathy vocal quality might be a useful compensatory strategy for reducing the perceptual severity of modulation of voice for individuals with tremor affecting the larynx.

  10. Participatory Investigation of the Great East Japan Disaster: PhotoVoice from Women Affected by the Calamity.

    PubMed

    Yoshihama, Mieko; Yunomae, Tomoko

    2018-05-02

    Disasters exacerbate predisaster inequities and intensify the vulnerability of women and other marginalized and disempowered groups. Thus, disaster policies and responses should incorporate the experiences and perspectives of those who are marginalized. The authors sought to conduct a participatory research project to help develop more inclusive, gender-informed disaster responses and policies in Japan. In June 2011, following three months of planning and preparation, they initiated a participatory examination of the impact of the Great East Japan Disaster using PhotoVoice methodology. Engaging the very women affected by the calamity, the authors first implemented the project in three localities in the hardest-hit areas of northern Japan-the prefectures of Fukushima, Miyagi, and Iwate. The authors have since expanded the project to other locations, and the project is ongoing. Focused on the planning, implementation, and outcomes of the initial phase, this article examines the role and potential of participatory action research using the PhotoVoice methodology in the aftermath of a major disaster.

  11. Numerical analysis of effects of transglottal pressure change on fundamental frequency of phonation.

    PubMed

    Deguchi, Shinji; Matsuzaki, Yuji; Ikeda, Tadashige

    2007-02-01

    In humans, a decrease in transglottal pressure (Pt) causes an increase in the fundamental frequency of phonation (F0) only at a specific voice pitch within the modal register, the mechanism of which remains unclear. In the present study, numerical analyses were performed to investigate the mechanism of the voice pitch-dependent positive change of F0 due to Pt decrease. The airflow and the airway, including the vocal folds, were modeled in terms of mechanics of fluid and structure. Simulations of phonation using the numerical model indicated that Pt affects both the average position and the average amplitude magnitude of vocal fold self-excited oscillation in a non-monotonous manner. This effect results in voice pitch-dependent responses of F0 to Pt decreases, including the positive response of F0 as actually observed in humans. The findings of the present study highlight the importance of considering self-excited oscillation of the vocal folds in elucidation of the phonation mechanism.

  12. Visual attention modulates brain activation to angry voices.

    PubMed

    Mothes-Lasch, Martin; Mentzel, Hans-Joachim; Miltner, Wolfgang H R; Straube, Thomas

    2011-06-29

    In accordance with influential models proposing prioritized processing of threat, previous studies have shown automatic brain responses to angry prosody in the amygdala and the auditory cortex under auditory distraction conditions. However, it is unknown whether the automatic processing of angry prosody is also observed during cross-modal distraction. The current fMRI study investigated brain responses to angry versus neutral prosodic stimuli during visual distraction. During scanning, participants were exposed to angry or neutral prosodic stimuli while visual symbols were displayed simultaneously. By means of task requirements, participants either attended to the voices or to the visual stimuli. While the auditory task revealed pronounced activation in the auditory cortex and amygdala to angry versus neutral prosody, this effect was absent during the visual task. Thus, our results show a limitation of the automaticity of the activation of the amygdala and auditory cortex to angry prosody. The activation of these areas to threat-related voices depends on modality-specific attention.

  13. Responsive Consumerism: Empowerment in Markets for Health Plans

    PubMed Central

    Elbel, Brian; Schlesinger, Mark

    2009-01-01

    Context: American health policy is increasingly relying on consumerism to improve its performance. This article examines a neglected aspect of medical consumerism: the extent to which consumers respond to problems with their health plans. Methods: Using a telephone survey of five thousand consumers conducted in 2002, this article assesses how frequently consumers voice formal grievances or exit from their health plan in response to problems of differing severity. This article also examines the potential impact of this responsiveness on both individuals and the market. In addition, using cross-group comparisons of means and regressions, it looks at how the responses of “empowered” consumers compared with those who are “less empowered.” Findings: The vast majority of consumers do not formally voice their complaints or exit health plans, even in response to problems with significant consequences. “Empowered” consumers are only minimally more likely to formally voice and no more likely to leave their plan. Moreover, given the greater prevalence of trivial problems, consumers are much more likely to complain or leave their plans because of problems that are not severe. Greater empowerment does not alleviate this. Conclusions: While much of the attention on consumerism has focused on prospective choice, understanding how consumers respond to problems is equally, if not more, important. Relying on consumers’ responses as a means to protect individual consumers or influence the market for health plans is unlikely to be successful in its current form. PMID:19751285

  14. An open-label study of sodium oxybate in Spasmodic dysphonia.

    PubMed

    Rumbach, Anna F; Blitzer, Andrew; Frucht, Steven J; Simonyan, Kristina

    2017-06-01

    Spasmodic dysphonia (SD) is a task-specific laryngeal dystonia that affects speech production. Co-occurring voice tremor (VT) often complicates the diagnosis and clinical management of SD. Treatment of SD and VT is largely limited to botulinum toxin injections into laryngeal musculature; other pharmacological options are not sufficiently developed. Open-label study. We conducted an open-label study in 23 SD and 22 SD/VT patients to examine the effects of sodium oxybate (Xyrem), an oral agent with therapeutic effects similar to those of alcohol in these patients. Blinded randomized analysis of voice and speech samples assessed symptom improvement before and after drug administration. Sodium oxybate significantly improved voice symptoms (P = .001) primarily by reducing the number of SD-characteristic voice breaks and severity of VT. Sodium oxybate further showed a trend for improving VT symptoms (P = .03) in a subset of patients who received successful botulinum toxin injections for the management of their SD symptoms. The drug's effects were observed approximately 30 to 40 minutes after its intake and lasted about 3.5 to 4 hours. Our study demonstrated that sodium oxybate reduced voice symptoms in 82.2% of alcohol-responsive SD patients both with and without co-occurring VT. Our findings suggest that the therapeutic mechanism of sodium oxybate in SD and SD/VT may be linked to that of alcohol, and as such, sodium oxybate might be beneficial for alcohol-responsive SD and SD/VT patients. 4 Laryngoscope, 127:1402-1407, 2017. © 2016 The American Laryngological, Rhinological and Otological Society, Inc.

  15. Chiropractic Care for a Patient with Spasmodic Dysphonia Associated with Cervical Spine Trauma

    PubMed Central

    Waddell, Roger K.

    2005-01-01

    Abstract Objective To discuss the diagnosis and response to treatment of spasmodic dysphonia in a 25-year-old female vocalist following an auto accident. Clinical Features The voice disorder and neck pain appeared after the traumatic incident. Examination of the cervical spine revealed moderate pain, muscle spasm and restricted joint motion at C-1 and C-5 on the left side. Cervical range of motion was reduced on left rotation. Bilateral manual muscle testing of the trapezius and sternocleidomastoid muscles, which share innervation with the laryngeal muscles by way of the spinal accessory nerve, were weak on the left side. Pre and post accident voice range profiles (phonetograms) that measure singing voice quality were examined. The pre- and post-accident phonetograms revealed significant reduction in voice intensity and fundamental frequency as measured in decibels and hertz. Intervention and Outcome Low-force chiropractic spinal manipulative therapy to C-1 and C-5 was employed. Following a course of care, the patient's singing voice returned to normal, as well as a resolution of her musculo- skeletal complaints. Conclusion It appears that in certain cases, the singing voice can be adversely affected if neck or head trauma is severe enough. This case proposes that trauma with irritation to the cervical spine nerve roots as they communicate with the spinal accessory, and in turn the laryngeal nerves, may be contributory in some functional voice disorders or muscle tension dysphonia. PMID:19674642

  16. ERP correlates of motivating voices: quality of motivation and time-course matters

    PubMed Central

    Zougkou, Konstantina; Weinstein, Netta

    2017-01-01

    Abstract Here, we conducted the first study to explore how motivations expressed through speech are processed in real-time. Participants listened to sentences spoken in two types of well-studied motivational tones (autonomy-supportive and controlling), or a neutral tone of voice. To examine this, listeners were presented with sentences that either signaled motivations through prosody (tone of voice) and words simultaneously (e.g. ‘You absolutely have to do it my way’ spoken in a controlling tone of voice), or lacked motivationally biasing words (e.g. ‘Why don’t we meet again tomorrow’ spoken in a motivational tone of voice). Event-related brain potentials (ERPs) in response to motivations conveyed through words and prosody showed that listeners rapidly distinguished between motivations and neutral forms of communication as shown in enhanced P2 amplitudes in response to motivational when compared with neutral speech. This early detection mechanism is argued to help determine the importance of incoming information. Once assessed, motivational language is continuously monitored and thoroughly evaluated. When compared with neutral speech, listening to controlling (but not autonomy-supportive) speech led to enhanced late potential ERP mean amplitudes, suggesting that listeners are particularly attuned to controlling messages. The importance of controlling motivation for listeners is mirrored in effects observed for motivations expressed through prosody only. Here, an early rapid appraisal, as reflected in enhanced P2 amplitudes, is only found for sentences spoken in controlling (but not autonomy-supportive) prosody. Once identified as sounding pressuring, the message seems to be preferentially processed, as shown by enhanced late potential amplitudes in response to controlling prosody. Taken together, results suggest that motivational and neutral language are differentially processed; further, the data suggest that listening to cues signaling pressure and control cannot be ignored and lead to preferential, and more in-depth processing mechanisms. PMID:28525641

  17. ERP correlates of motivating voices: quality of motivation and time-course matters.

    PubMed

    Zougkou, Konstantina; Weinstein, Netta; Paulmann, Silke

    2017-10-01

    Here, we conducted the first study to explore how motivations expressed through speech are processed in real-time. Participants listened to sentences spoken in two types of well-studied motivational tones (autonomy-supportive and controlling), or a neutral tone of voice. To examine this, listeners were presented with sentences that either signaled motivations through prosody (tone of voice) and words simultaneously (e.g. 'You absolutely have to do it my way' spoken in a controlling tone of voice), or lacked motivationally biasing words (e.g. 'Why don't we meet again tomorrow' spoken in a motivational tone of voice). Event-related brain potentials (ERPs) in response to motivations conveyed through words and prosody showed that listeners rapidly distinguished between motivations and neutral forms of communication as shown in enhanced P2 amplitudes in response to motivational when compared with neutral speech. This early detection mechanism is argued to help determine the importance of incoming information. Once assessed, motivational language is continuously monitored and thoroughly evaluated. When compared with neutral speech, listening to controlling (but not autonomy-supportive) speech led to enhanced late potential ERP mean amplitudes, suggesting that listeners are particularly attuned to controlling messages. The importance of controlling motivation for listeners is mirrored in effects observed for motivations expressed through prosody only. Here, an early rapid appraisal, as reflected in enhanced P2 amplitudes, is only found for sentences spoken in controlling (but not autonomy-supportive) prosody. Once identified as sounding pressuring, the message seems to be preferentially processed, as shown by enhanced late potential amplitudes in response to controlling prosody. Taken together, results suggest that motivational and neutral language are differentially processed; further, the data suggest that listening to cues signaling pressure and control cannot be ignored and lead to preferential, and more in-depth processing mechanisms. © The Author (2017). Published by Oxford University Press.

  18. Mouth and Voice: A Relationship between Visual and Auditory Preference in the Human Superior Temporal Sulcus

    PubMed Central

    2017-01-01

    Cortex in and around the human posterior superior temporal sulcus (pSTS) is known to be critical for speech perception. The pSTS responds to both the visual modality (especially biological motion) and the auditory modality (especially human voices). Using fMRI in single subjects with no spatial smoothing, we show that visual and auditory selectivity are linked. Regions of the pSTS were identified that preferred visually presented moving mouths (presented in isolation or as part of a whole face) or moving eyes. Mouth-preferring regions responded strongly to voices and showed a significant preference for vocal compared with nonvocal sounds. In contrast, eye-preferring regions did not respond to either vocal or nonvocal sounds. The converse was also true: regions of the pSTS that showed a significant response to speech or preferred vocal to nonvocal sounds responded more strongly to visually presented mouths than eyes. These findings can be explained by environmental statistics. In natural environments, humans see visual mouth movements at the same time as they hear voices, while there is no auditory accompaniment to visual eye movements. The strength of a voxel's preference for visual mouth movements was strongly correlated with the magnitude of its auditory speech response and its preference for vocal sounds, suggesting that visual and auditory speech features are coded together in small populations of neurons within the pSTS. SIGNIFICANCE STATEMENT Humans interacting face to face make use of auditory cues from the talker's voice and visual cues from the talker's mouth to understand speech. The human posterior superior temporal sulcus (pSTS), a brain region known to be important for speech perception, is complex, with some regions responding to specific visual stimuli and others to specific auditory stimuli. Using BOLD fMRI, we show that the natural statistics of human speech, in which voices co-occur with mouth movements, are reflected in the neural architecture of the pSTS. Different pSTS regions prefer visually presented faces containing either a moving mouth or moving eyes, but only mouth-preferring regions respond strongly to voices. PMID:28179553

  19. Mouth and Voice: A Relationship between Visual and Auditory Preference in the Human Superior Temporal Sulcus.

    PubMed

    Zhu, Lin L; Beauchamp, Michael S

    2017-03-08

    Cortex in and around the human posterior superior temporal sulcus (pSTS) is known to be critical for speech perception. The pSTS responds to both the visual modality (especially biological motion) and the auditory modality (especially human voices). Using fMRI in single subjects with no spatial smoothing, we show that visual and auditory selectivity are linked. Regions of the pSTS were identified that preferred visually presented moving mouths (presented in isolation or as part of a whole face) or moving eyes. Mouth-preferring regions responded strongly to voices and showed a significant preference for vocal compared with nonvocal sounds. In contrast, eye-preferring regions did not respond to either vocal or nonvocal sounds. The converse was also true: regions of the pSTS that showed a significant response to speech or preferred vocal to nonvocal sounds responded more strongly to visually presented mouths than eyes. These findings can be explained by environmental statistics. In natural environments, humans see visual mouth movements at the same time as they hear voices, while there is no auditory accompaniment to visual eye movements. The strength of a voxel's preference for visual mouth movements was strongly correlated with the magnitude of its auditory speech response and its preference for vocal sounds, suggesting that visual and auditory speech features are coded together in small populations of neurons within the pSTS. SIGNIFICANCE STATEMENT Humans interacting face to face make use of auditory cues from the talker's voice and visual cues from the talker's mouth to understand speech. The human posterior superior temporal sulcus (pSTS), a brain region known to be important for speech perception, is complex, with some regions responding to specific visual stimuli and others to specific auditory stimuli. Using BOLD fMRI, we show that the natural statistics of human speech, in which voices co-occur with mouth movements, are reflected in the neural architecture of the pSTS. Different pSTS regions prefer visually presented faces containing either a moving mouth or moving eyes, but only mouth-preferring regions respond strongly to voices. Copyright © 2017 the authors 0270-6474/17/372697-12$15.00/0.

  20. Cognitive Load in Voice Therapy Carry-Over Exercises.

    PubMed

    Iwarsson, Jenny; Morris, David Jackson; Balling, Laura Winther

    2017-01-01

    The cognitive load generated by online speech production may vary with the nature of the speech task. This article examines 3 speech tasks used in voice therapy carry-over exercises, in which a patient is required to adopt and automatize new voice behaviors, ultimately in daily spontaneous communication. Twelve subjects produced speech in 3 conditions: rote speech (weekdays), sentences in a set form, and semispontaneous speech. Subjects simultaneously performed a secondary visual discrimination task for which response times were measured. On completion of each speech task, subjects rated their experience on a questionnaire. Response times from the secondary, visual task were found to be shortest for the rote speech, longer for the semispontaneous speech, and longest for the sentences within the set framework. Principal components derived from the subjective ratings were found to be linked to response times on the secondary visual task. Acoustic measures reflecting fundamental frequency distribution and vocal fold compression varied across the speech tasks. The results indicate that consideration should be given to the selection of speech tasks during the process leading to automation of revised speech behavior and that self-reports may be a reliable index of cognitive load.

  1. Utilization of Internet Protocol-Based Voice Systems in Remote Payload Operations

    NASA Technical Reports Server (NTRS)

    Chamberlain, jim; Bradford, Bob; Best, Susan; Nichols, Kelvin

    2002-01-01

    Due to limited crew availability to support science and the large number of experiments to be operated simultaneously, telescience is key to a successful International Space Station (ISS) science program. Crew, operations personnel at NASA centers, and researchers at universities and companies around the world must work closely together to per orm scientific experiments on-board ISS. The deployment of reliable high-speed Internet Protocol (IP)-based networks promises to greatly enhance telescience capabilities. These networks are now being used to cost-effectively extend the reach of remote mission support systems. They reduce the need for dedicated leased lines and travel while improving distributed workgroup collaboration capabilities. NASA has initiated use of Voice over Internet Protocol (VoIP) to supplement the existing mission voice communications system used by researchers at their remote sites. The Internet Voice Distribution System (IVoDS) connects remote researchers to mission support "loopsll or conferences via NASA networks and Internet 2. Researchers use NODS software on personal computers to talk with operations personnel at NASA centers. IVoDS also has the ;capability, if authorized, to allow researchers to communicate with the ISS crew during experiment operations. NODS was developed by Marshall Space Flight Center with contractors & Technology, First Virtual Communications, Lockheed-Martin, and VoIP Group. NODS is currently undergoing field-testing with full deployment for up to 50 simultaneous users expected in 2002. Research is being performed in parallel with IVoDS deployment for a next-generation system to qualitatively enhance communications among ISS operations personnel. In addition to the current voice capability, video and data/application-sharing capabilities are being investigated. IVoDS technology is also being considered for mission support systems for programs such as Space Launch Initiative and Homeland Defense.

  2. The Federal Telecommunications System 2000, a Military Perspective

    DTIC Science & Technology

    1988-01-01

    w . - I LL) p SECURITY CLASSIFICATIOr; C," THIS PAGE (When Data Ftvered) ’ J, REPORT DOCUMENTATION PACE READ INSTRUINT, R•PORT NUMBER iT ACCESSION No...the federal government’s telecommunications problems. This system will offer voice, data and video services across a transparent, nationwide network...of fiscal realities, proposes the FTS2000 as the answer to the federal government’s telecommunications problems. This system will * offer voice, data

  3. The Effect of Hydration on Voice Quality in Adults: A Systematic Review.

    PubMed

    Alves, Maxine; Krüger, Esedra; Pillay, Bhavani; van Lierde, Kristiane; van der Linde, Jeannie

    2017-11-06

    We aimed to critically appraise scientific, peer-reviewed articles, published in the past 10 years on the effects of hydration on voice quality in adults. This is a systematic review. Five databases were searched using the key words "vocal fold hydration", "voice quality", "vocal fold dehydration", and "hygienic voice therapy". The Preferred Reporting Items for Systematic Review and Meta-Analyses (PRISMA) guidelines were followed. The included studies were scored based on American Speech-Language-Hearing Association's levels of evidence and quality indicators, as well as the Cochrane Collaboration's risk of bias tool. Systemic dehydration as a result of fasting and not ingesting fluids significantly negatively affected the parameters of noise-to-harmonics ratio (NHR), shimmer, jitter, frequency, and the s/z ratio. Water ingestion led to significant improvements in shimmer, jitter, frequency, and maximum phonation time values. Caffeine intake does not appear to negatively affect voice production. Laryngeal desiccation challenges by oral breathing led to surface dehydration which negatively affected jitter, shimmer, NHR, phonation threshold pressure, and perceived phonatory effort. Steam inhalation significantly improved NHR, shimmer, and jitter. Only nebulization of isotonic solution decreased phonation threshold pressure and showed some indication of a potential positive effect of nebulization substances. Treatments in high humidity environments prove to be effective and adaptations of low humidity environments should be encouraged. Recent literature regarding vocal hydration is high quality evidence. Systemic hydration is the easiest and most cost-effective solution to improve voice quality. Recent evidence therefore supports the inclusion of hydration in a vocal hygiene program. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  4. Ad spending: maintaining market share.

    PubMed

    Jones, J P

    1990-01-01

    Accuracy in manufacturers' advertising budgeting is hampered by reliance on the case rate system, which ties budgets to sales. A better measure is a brand's market share compared with its share of voice (the brand's share of the total value of the main media exposure in that product category). New brands are often "investing" in the market: speaking in a louder voice than their market shares would justify. Popular brands are often "profit taking"--keeping their voices low but enjoying a disproportionately large market share. The interrelationship between market share and share of voice, with either "investing" or "profit taking" the desired result, is not usually considered when determining ad budgets. But as advertisers realize how market share can respond to advertising pressure through switches in the share of voice, this method of market testing should gain in importance.

  5. Effects of audio compression in automatic detection of voice pathologies.

    PubMed

    Sáenz-Lechón, Nicolás; Osma-Ruiz, Víctor; Godino-Llorente, Juan I; Blanco-Velasco, Manuel; Cruz-Roldán, Fernando; Arias-Londoño, Julián D

    2008-12-01

    This paper investigates the performance of an automatic system for voice pathology detection when the voice samples have been compressed in MP3 format and different binary rates (160, 96, 64, 48, 24, and 8 kb/s). The detectors employ cepstral and noise measurements, along with their derivatives, to characterize the voice signals. The classification is performed using Gaussian mixtures models and support vector machines. The results between the different proposed detectors are compared by means of detector error tradeoff (DET) and receiver operating characteristic (ROC) curves, concluding that there are no significant differences in the performance of the detector when the binary rates of the compressed data are above 64 kb/s. This has useful applications in telemedicine, reducing the storage space of voice recordings or transmitting them over narrow-band communications channels.

  6. Some effects of stress on users of a voice recognition system: A preliminary inquiry

    NASA Astrophysics Data System (ADS)

    French, B. A.

    1983-03-01

    Recent work with Automatic Speech Recognition has focused on applications and productivity considerations in the man-machine interface. This thesis is an attempt to see if placing users of such equipment under time-induced stress has an effect on their percent correct recognition rates. Subjects were given a message-handling task of fixed length and allowed progressively shorter times to attempt to complete it. Questionnaire responses indicate stress levels increased with decreased time-allowance; recognition rates decreased as time was reduced.

  7. Engineering Evaluation and Assessment (EE and A) Report for the Symbolic and Sub-symbolic Robotics Intelligence Control System (SS-RICS)

    DTIC Science & Technology

    2018-04-01

    Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions...2006. Since that time , SS-RICS has been the integration platform for many robotics algorithms using a variety of different disciplines from cognitive...voice recognition. Each noise level was run 10 times per gender, yielding 60 total runs. Two paths were chosen for testing (Paths A and B) of

  8. Some Effects of Stress on Users of a Voice Recognition System: A Preliminary Inquiry.

    DTIC Science & Technology

    1983-03-01

    criterion of face valiaity Is also imposed (i.e., tne tasks are cctiigrea tc be acce;table to ta:get populations, e.g., pilots ... .Re :.11: pp. 22-Z5j...being the hlgtesz level of eacrt. it was thougtt tnat these iraividual response levels trighz soirehow te reiated to recognitio , rates. L. CUNCI-kTAL...generalizable ;heuorenon, it would irply that after some few training sessions with a reccenizer, the distinction vanishes. If so, faced with a

  9. Task-specific singing dystonia: vocal instability that technique cannot fix.

    PubMed

    Halstead, Lucinda A; McBroom, Deanna M; Bonilha, Heather Shaw

    2015-01-01

    Singer's dystonia is a rare variation of focal laryngeal dystonia presenting only during specific tasks in the singing voice. It is underdiagnosed since it is commonly attributed to technique problems including increased muscle tension, register transition, or wobble. Singer's dystonia differs from technique-related issues in that it is task- and/or pitch-specific, reproducible and occurs independently from the previously mentioned technical issues.This case series compares and contrasts profiles of four patients with singer's dystonia to increase our knowledge of this disorder. This retrospective case series includes a detailed case history, results of singing evaluations from individual voice teachers, review of singing voice samples by a singing voice specialist, evaluation by a laryngologist with endoscopy and laryngeal electromyography (LEMG), and spectral analysis of the voice samples by a speech-language pathologist. Results demonstrate the similarities and unique differences of individuals with singer's dystonia. Response to treatment and singing status varied from nearly complete relief of symptoms with botulinum toxin injections to minor relief of symptoms and discontinuation of singing. The following are the conclusions from this case series: (1) singer's dystonia exists as a separate entity from technique issues, (2) singer's dystonia is consistent with other focal task-specific dystonias found in musicians, (3) correctly diagnosing singer's dystonia allows singer's access to medical treatment of dystonia and an opportunity to modify their singing repertoire to continue singing with the voice they have, and (4) diagnosis of singer's dystonia requires careful sequential multidisciplinary evaluation to isolate the instability and confirm dystonia by LEMG and spectral voice analysis. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  10. Development and Validation of the Children's Voice Handicap Index-10 for Parents.

    PubMed

    Ricci-Maccarini, Andrea; De Maio, Vincenzo; Murry, Thomas; Schindler, Antonio

    2016-01-01

    The Children's Voice Handicap Index-10 (CVHI-10) was introduced as a tool for self-assessment of children's dysphonia. However, in the management of children with voice disorders, both parents' and children's perspectives play an important role. Because a self-tool including both a children's and a parents' version does not exist yet, the aim of the study was to develop and validate an assessment tool which parallels the CVHI-10 for parents to assess the level of voice handicap in their child's voice. Observational, prospective, cross-sectional study. To develop a CVHI-10 for parents, called "CVHI-10-P", the CVHI-10 items were adapted to reflect parents' responses about their child. Fifty-five children aged 7-12 years completed the CVHI-10, whereas their parents completed the CVHI-10-P. Each child's voice was also perceptually assessed by an otolaryngologist using the Grade Breathness Roughness (GRB) scale. Fifty-one of the 55 children underwent voice therapy (VT) and were assessed afterward using the GRB, CVHI-10, and CVHI-10-P. CVHI-10-P internal consistency was satisfactory (Cronbach alpha = .78). Correlation between CVHI-10-P and CVHI-10 was moderate (r = 0.37). CVHI-10-P total scores were lower than CVHI-10 scores in most of the cases. Single-item mean scores were always lower in CVHI-10-P compared with CVHI-10, with the exception of the only one item of the CVHI-10-P that directly involves the parent's experience (item 10). Data gained from one tool are not directly related to the other, suggesting that these two tools appraise the child's voice handicap from different perspectives. The overall perceptual assessment scores of the 51 children after VT significantly improved. There was a statistically significant reduction of the total scores and for each item in CVHI-10 and CVHI-10-P after VT. These data support the adoption of the CVHI-10-P as an assessment tool and an outcome measure for management of children's voice disorders. CVHI-10-P is a valid tool to appraise parents' perspective of their child's voice disorder. The use of the CVHI-10 and the CVHI-10-P is recommended for objectively determining the level of voice handicap in children by parents and child. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  11. Study of Risk Factors for Development of Voice Disorders and its Impact on the Quality of Life of School Teachers in Mangalore, India.

    PubMed

    Alva, Arati; Machado, Megna; Bhojwani, Kiran; Sreedharan, Suja

    2017-01-01

    School teachers are most prone to the development and detrimental effects of voice disorders as a consequence of their work. The risk factors for development of dysphonia in teachers are multifactorial. The primary aim of our study was to investigate the various risk factors that influence the onset and progression of voice disorders in school teachers in the Indian context. We wanted to assess the effect of voice problems on the physical, psychosocial and functional aspect of a teacher's life. It was a cross-sectional study conducted across three English medium institutions. A total of 105 teachers consented to participate in the study and they had to answer a semi-structured, pre-tested questionnaire, which included demographic details, living habits (drug intake, smoking and alcohol intake) health condition [any Deviated Nasal Septum (DNS), Gastroesophageal Reflux Disease (GERD), stress, etc., or any history of surgery], teaching characteristics, voice symptoms and physical discomforts and quality of life assessment. The completed questionnaires were collected and analyzed based on the responses obtained. It was found that 81% of the study population had voice problems at some point of their career. A total of 26% of them fell into the voice disorder category. The association of upper respiratory infections DNS and GERD with voice disorders was found to be statistically significant. We also found that a significant number of teachers with voice disorders had changed their teaching styles and were planning to opt for an early retirement. Most importantly, it was also seen that teachers with voice disorders were more likely to have a poorer quality of life as compared to those without a voice disorder (p<0.001). Voice disorders had a significant bearing on all the spheres of a school teacher's life. The affected teachers were more likely to take sick leaves, change overall job opinions, retire early, reduce overall communiations, repeat statements and avoid talking to people in person as well as over the telephone. It reduced their overall social abilities and made them avoid social activities. They got easily upset and were dissatisfied with their job performance. All these in turn deteriorate the quality of life in these individuals.

  12. Study of Risk Factors for Development of Voice Disorders and its Impact on the Quality of Life of School Teachers in Mangalore, India

    PubMed Central

    Machado, Megna; Bhojwani, Kiran; Sreedharan, Suja

    2017-01-01

    Introduction School teachers are most prone to the development and detrimental effects of voice disorders as a consequence of their work. The risk factors for development of dysphonia in teachers are multifactorial. Aim The primary aim of our study was to investigate the various risk factors that influence the onset and progression of voice disorders in school teachers in the Indian context. We wanted to assess the effect of voice problems on the physical, psychosocial and functional aspect of a teacher’s life. Materials and Methods It was a cross-sectional study conducted across three English medium institutions. A total of 105 teachers consented to participate in the study and they had to answer a semi-structured, pre-tested questionnaire, which included demographic details, living habits (drug intake, smoking and alcohol intake) health condition [any Deviated Nasal Septum (DNS), Gastroesophageal Reflux Disease (GERD), stress, etc., or any history of surgery], teaching characteristics, voice symptoms and physical discomforts and quality of life assessment. The completed questionnaires were collected and analyzed based on the responses obtained. Results It was found that 81% of the study population had voice problems at some point of their career. A total of 26% of them fell into the voice disorder category. The association of upper respiratory infections DNS and GERD with voice disorders was found to be statistically significant. We also found that a significant number of teachers with voice disorders had changed their teaching styles and were planning to opt for an early retirement. Most importantly, it was also seen that teachers with voice disorders were more likely to have a poorer quality of life as compared to those without a voice disorder (p<0.001). Conclusion Voice disorders had a significant bearing on all the spheres of a school teacher’s life. The affected teachers were more likely to take sick leaves, change overall job opinions, retire early, reduce overall communiations, repeat statements and avoid talking to people in person as well as over the telephone. It reduced their overall social abilities and made them avoid social activities. They got easily upset and were dissatisfied with their job performance. All these in turn deteriorate the quality of life in these individuals. PMID:28273984

  13. Differential neural contributions to native- and foreign-language talker identification

    PubMed Central

    Perrachione, Tyler K.; Pierrehumbert, Janet B.; Wong, Patrick C.M.

    2009-01-01

    Humans are remarkably adept at identifying individuals by the sound of their voice, a behavior supported by the nervous system’s ability to integrate information from voice and speech perception. Talker-identification abilities are significantly impaired when listeners are unfamiliar with the language being spoken. Recent behavioral studies describing the language-familiarity effect implicate functionally integrated neural systems for speech and voice perception, yet specific neuroscientific evidence demonstrating the basis for such integration has not yet been shown. Listeners in the present study learned to identify voices speaking a familiar (native) or unfamiliar (foreign) language. The talker-identification performance of neural circuitry in each cerebral hemisphere was assessed using dichotic listening. To determine the relative contribution of circuitry in each hemisphere to ecological (binaural) talker identification abilities, we compared the predictive capacity of dichotic performance on binaural performance across languages. We found listeners’ right-ear (left hemisphere) performance to be a better predictor of overall accuracy in their native language than a foreign one. The enhanced predictive capacity of the classically language-dominant left-hemisphere on overall talker-identification accuracy demonstrates functionally integrated neural systems for speech and voice perception during natural talker identification. PMID:19968445

  14. The effect of deep brain stimulation on the speech motor system.

    PubMed

    Mücke, Doris; Becker, Johannes; Barbe, Michael T; Meister, Ingo; Liebhart, Lena; Roettger, Timo B; Dembek, Till; Timmermann, Lars; Grice, Martine

    2014-08-01

    Chronic deep brain stimulation of the nucleus ventralis intermedius is an effective treatment for individuals with medication-resistant essential tremor. However, these individuals report that stimulation has a deleterious effect on their speech. The present study investigates one important factor leading to these effects: the coordination of oral and glottal articulation. Sixteen native-speaking German adults with essential tremor, between 26 and 86 years old, with and without chronic deep brain stimulation of the nucleus ventralis intermedius and 12 healthy, age-matched subjects were recorded performing a fast syllable repetition task (/papapa/, /tatata/, /kakaka/). Syllable duration and voicing-to-syllable ratio as well as parameters related directly to consonant production, voicing during constriction, and frication during constriction were measured. Voicing during constriction was greater in subjects with essential tremor than in controls, indicating a perseveration of voicing into the voiceless consonant. Stimulation led to fewer voiceless intervals (voicing-to-syllable ratio), indicating a reduced degree of glottal abduction during the entire syllable cycle. Stimulation also induced incomplete oral closures (frication during constriction), indicating imprecise oral articulation. The detrimental effect of stimulation on the speech motor system can be quantified using acoustic measures at the subsyllabic level.

  15. Mapping Phonetic Features for Voice-Driven Sound Synthesis

    NASA Astrophysics Data System (ADS)

    Janer, Jordi; Maestre, Esteban

    In applications where the human voice controls the synthesis of musical instruments sounds, phonetics convey musical information that might be related to the sound of the imitated musical instrument. Our initial hypothesis is that phonetics are user- and instrument-dependent, but they remain constant for a single subject and instrument. We propose a user-adapted system, where mappings from voice features to synthesis parameters depend on how subjects sing musical articulations, i.e. note to note transitions. The system consists of two components. First, a voice signal segmentation module that automatically determines note-to-note transitions. Second, a classifier that determines the type of musical articulation for each transition based on a set of phonetic features. For validating our hypothesis, we run an experiment where subjects imitated real instrument recordings with their voice. Performance recordings consisted of short phrases of saxophone and violin performed in three grades of musical articulation labeled as: staccato, normal, legato. The results of a supervised training classifier (user-dependent) are compared to a classifier based on heuristic rules (user-independent). Finally, from the previous results we show how to control the articulation in a sample-concatenation synthesizer by selecting the most appropriate samples.

  16. The effect of processing code, response modality and task difficulty on dual task performance and subjective workload in a manual system

    NASA Technical Reports Server (NTRS)

    Liu, Yili; Wickens, Christopher D.

    1987-01-01

    This paper reports on the first experiment of a series studying the effect of task structure and difficulty demand on time-sharing performance and workload in both automated and corresponding manual systems. The experimental task involves manual control time-shared with spatial and verbal decisions tasks of two levels of difficulty and two modes of response (voice or manual). The results provide strong evidence that tasks and processes competing for common processing resources are time shared less effecively and have higher workload than tasks competing for separate resources. Subjective measures and the structure of multiple resources are used in conjunction to predict dual task performance. The evidence comes from both single-task and from dual-task performance.

  17. Communication strategies and timeliness of response to life critical telemetry alarms.

    PubMed

    Bonzheim, Kimberly A; Gebara, Rani I; O'Hare, Bridget M; Ellis, R Darin; Brand, Monique A; Balar, Salil D; Stockman, Rita; Sciberras, Annette M; Haines, David E

    2011-05-01

    A centralized electrocardiogram telemetry monitoring system (TMS) facilitates early identification of critical arrhythmias and acute medical decompensation. Timely intervention can only be performed if abnormalities are communicated rapidly to the direct caregiver. The study objectives were to measure effectiveness of bi-directional voice communication badges versus one-way alphanumeric pagers for telemetry alarm response and communication loop closure. A sequential observational pilot study of nursing response to TMS alarms compared communication technologies on four nursing units in a 1,061 bed tertiary care hospital with 264 TMS channels of telemetry over a 2-year period. Subsequently, the communication technologies were compared in a randomized fashion on a 68-bed progressive cardiac care unit. Caregivers were blinded to the protocol. All alarm responses were recorded during two periods using either pagers or voice communication devices. Alarm response time and closure of the communication loop were analyzed in a blinded fashion. The direct communication functionality of the badge significantly shortened the time to first contact, time to completion, and rate of closure of the communication loop in both the pilot and study phases. Median time to first contact with the communication badge was 0.5  min, compared to 1.6  min with pager communication (p < 0.0003). Communication loop closure was achieved in 100% of clinical alarms using the badge versus 19% with the pager (p < 0.0001). Communication badge technology reduced alarm time to first contact and completion as well as facilitated communication loop closures. Immediate two-way communication significantly impacted practice, alarm management, and resulted in faster bedside care.

  18. PACS administrators' and radiologists' perspective on the importance of features for PACS selection.

    PubMed

    Joshi, Vivek; Narra, Vamsi R; Joshi, Kailash; Lee, Kyootai; Melson, David

    2014-08-01

    Picture archiving and communication systems (PACS) play a critical role in radiology. This paper presents the criteria important to PACS administrators for selecting a PACS. A set of criteria are identified and organized into an integrative hierarchical framework. Survey responses from 48 administrators are used to identify the relative weights of these criteria through an analytical hierarchy process. The five main dimensions for PACS selection in order of importance are system continuity and functionality, system performance and architecture, user interface for workflow management, user interface for image manipulation, and display quality. Among the subdimensions, the highest weights were assessed for security, backup, and continuity; tools for continuous performance monitoring; support for multispecialty images; and voice recognition/transcription. PACS administrators' preferences were generally in line with that of previously reported results for radiologists. Both groups assigned the highest priority to ensuring business continuity and preventing loss of data through features such as security, backup, downtime prevention, and tools for continuous PACS performance monitoring. PACS administrators' next high priorities were support for multispecialty images, image retrieval speeds from short-term and long-term storage, real-time monitoring, and architectural issues of compatibility and integration with other products. Thus, next to ensuring business continuity, administrators' focus was on issues that impact their ability to deliver services and support. On the other hand, radiologists gave high priorities to voice recognition, transcription, and reporting; structured reporting; and convenience and responsiveness in manipulation of images. Thus, radiologists' focus appears to be on issues that may impact their productivity, effort, and accuracy.

  19. The effect of voice communications latency in high density, communications-intensive airspace.

    DOT National Transportation Integrated Search

    2003-01-01

    The Federal Aviation Administration (FAA) Next Generation Air-Ground Communications program plans to replace aging analog radio equipment with the Very High Frequency Digital Link Mode 3 (VDL3) system. VDL3 will implement both digital voice and data ...

  20. Using Continuous Voice Recognition Technology as an Input Medium to the Naval Warfare Interactive Simulation System (NWISS).

    DTIC Science & Technology

    1984-06-01

    Co ,u’arataor, Gr 7- / ’ . c ; / , caae.ic >ar. ’ ’# d:.i II ’ ..... .. . . .. .. . ... . , rV ABSTRACT A great d-al of research has been conducted an...9 2. Continuous Voice -%ecoait.ior, ....... 11 B. VERBEX 3000 SPEECH APPLiCATION DEVELOP !ENT SYSTEM! ( SPADS ...13 C . NAVAL IAR FARE INT7EACTI7E S:AIULATIC"N SYSTEM (NWISS) ....... .................. 14 D. PURPOSE .................... 16 1. A Past

  1. An adaptive narrow band frequency modulation voice communication system

    NASA Technical Reports Server (NTRS)

    Wishna, S.

    1972-01-01

    A narrow band frequency modulation communication system is described which provides for the reception of good quality voice at low carrier-to-noise ratios. The high level of performance is obtained by designing a limiter and phase lock loop combination as a demodulator, so that the bandwidth of the phase lock loop decreases as the carrier level decreases. The system was built for the position location and aircraft communication equipment experiment of the ATS 6 program.

  2. Autonomic Nervous System Responses During Perception of Masked Speech may Reflect Constructs other than Subjective Listening Effort

    PubMed Central

    Francis, Alexander L.; MacPherson, Megan K.; Chandrasekaran, Bharath; Alvar, Ann M.

    2016-01-01

    Typically, understanding speech seems effortless and automatic. However, a variety of factors may, independently or interactively, make listening more effortful. Physiological measures may help to distinguish between the application of different cognitive mechanisms whose operation is perceived as effortful. In the present study, physiological and behavioral measures associated with task demand were collected along with behavioral measures of performance while participants listened to and repeated sentences. The goal was to measure psychophysiological reactivity associated with three degraded listening conditions, each of which differed in terms of the source of the difficulty (distortion, energetic masking, and informational masking), and therefore were expected to engage different cognitive mechanisms. These conditions were chosen to be matched for overall performance (keywords correct), and were compared to listening to unmasked speech produced by a natural voice. The three degraded conditions were: (1) Unmasked speech produced by a computer speech synthesizer, (2) Speech produced by a natural voice and masked byspeech-shaped noise and (3) Speech produced by a natural voice and masked by two-talker babble. Masked conditions were both presented at a -8 dB signal to noise ratio (SNR), a level shown in previous research to result in comparable levels of performance for these stimuli and maskers. Performance was measured in terms of proportion of key words identified correctly, and task demand or effort was quantified subjectively by self-report. Measures of psychophysiological reactivity included electrodermal (skin conductance) response frequency and amplitude, blood pulse amplitude and pulse rate. Results suggest that the two masked conditions evoked stronger psychophysiological reactivity than did the two unmasked conditions even when behavioral measures of listening performance and listeners’ subjective perception of task demand were comparable across the three degraded conditions. PMID:26973564

  3. Functional hoarseness in children: short-term play therapy with family dynamic counseling as therapy of choice.

    PubMed

    Kollbrunner, Jürg; Seifert, Eberhard

    2013-09-01

    Children with nonorganic voice disorders (NVDs) are treated mainly using direct voice therapy techniques such as the accent method or glottal attack changes and indirect methods such as vocal hygiene and voice education. However, both approaches tackle only the symptoms and not etiological factors in the family dynamics and therefore often enjoy little success. The aim of the "Bernese Brief Dynamic Intervention" (BBDI) for children with NVD was to extend the effectiveness of pediatric voice therapies with a psychosomatic concept combining short-term play therapy with the child and family dynamic counseling of the parents. This study compares the therapeutic changes in three groups where different procedures were used, before intervention and 1 year afterward: counseling of parents (one to two consultations; n = 24), Brief Dynamic Intervention on the lines of the BBDI (three to five play therapy sessions with the child plus two to four sessions with the parents; n = 20), and traditional voice therapy (n = 22). A Voice Questionnaire for Parents developed by us with 59 questions to be answered on a four-point Likert scale was used to measure the change. According to the parents' assessment, a significant improvement in voice quality was achieved in all three methods. Counseling of parents (A) appears to have led parents to give their child more latitude, for example, they stopped nagging the child or demanding that he/she should behave strictly by the rules. After BBDI (B), the mothers were more responsive to their children's wishes and the children were more relaxed and their speech became livelier. At home, they called out to them less often at a distance, which probably improved parent-child dialog. Traditional voice therapy (C) seems to have had a positive effect on the children's social competence. BBDI seems to have the deepest, widest, and therefore probably the most enduring therapeutic effect on children with NVD. Copyright © 2013 The Voice Foundation. Published by Mosby, Inc. All rights reserved.

  4. High-Bandwidth Tactical-Network Data Analysis in a High-Performance-Computing (HPC) Environment: Voice Call Analysis

    DTIC Science & Technology

    2015-09-01

    Gateway 2 4. Voice Packet Flow: SIP , Session Description Protocol (SDP), and RTP 3 5. Voice Data Analysis 5 6. Call Analysis 6 7. Call Metrics 6...analysis processing is designed for a general VoIP system architecture based on Session Initiation Protocol ( SIP ) for negotiating call sessions and...employs Skinny Client Control Protocol for network communication between the phone and the local CallManager (e.g., for each dialed digit), SIP

  5. Collecting Self-Reported Data on Dating Abuse Perpetration From a Sample of Primarily Black and Hispanic, Urban-Residing, Young Adults: A Comparison of Timeline Followback Interview and Interactive Voice Response Methods.

    PubMed

    Rothman, Emily F; Heeren, Timothy; Winter, Michael; Dorfman, David; Baughman, Allyson; Stuart, Gregory

    2016-12-01

    Dating abuse is a prevalent and consequential public health problem. However, relatively few studies have compared methods of collecting self-report data on dating abuse perpetration. This study compares two data collection methods-(a) the Timeline Followback (TLFB) retrospective reporting method, which makes use of a written calendar to prompt respondents' recall, and (b) an interactive voice response (IVR) system, which is a prospective telephone-based database system that necessitates respondents calling in and entering data using their telephone keypads. We collected 84 days of data on young adult dating abuse perpetration using IVR from a total of 60 respondents. Of these respondents, 41 (68%) completed a TLFB retrospective report pertaining to the same 84-day period after that time period had ended. A greater number of more severe dating abuse perpetration events were reported via the IVR system. Participants who reported any dating abuse perpetration were more likely to report more frequent abuse perpetration via the IVR than the TLFB (i.e., may have minimized the number of times they perpetrated dating abuse on the TLFB). The TLFB method did not result in a tapering off of reported events past the first week as it has in prior studies, but the IVR method did result in a tapering off of reported events after approximately the sixth week. We conclude that using an IVR system for self-reports of dating abuse perpetration may not have substantial advantages over using a TLFB method, but researchers' choice of mode may vary by research question, resources, sample, and setting.

  6. Female voice communications in high levels of aircraft cockpit noises--Part I: spectra, levels, and microphones.

    PubMed

    Nixon, C W; Morris, L J; McCavitt, A R; McKinley, R L; Anderson, T R; McDaniel, M P; Yeager, D G

    1998-07-01

    Female produced speech, although more intelligible than male speech in some noise spectra, may be more vulnerable to degradation by high levels of some military aircraft cockpit noises. The acoustic features of female speech are higher in frequency, lower in power, and appear more susceptible than male speech to masking by some of these military noises. Current military aircraft voice communication systems were optimized for the male voice and may not adequately accommodate the female voice in these high level noises. This applied study investigated the intelligibility of female and male speech produced in the noise spectra of four military aircraft cockpits at levels ranging from 95 dB to 115 dB. The experimental subjects used standard flight helmets and headsets, noise-canceling microphones, and military aircraft voice communications systems during the measurements. The intelligibility of female speech was lower than that of male speech for all experimental conditions; however, differences were small and insignificant except at the highest levels of the cockpit noises. Intelligibility for both genders varied with aircraft noise spectrum and level. Speech intelligibility of both genders was acceptable during normal cruise noises of all four aircraft, but improvements are required in the higher levels of noise created during aircraft maximum operating conditions. The intelligibility of female speech was unacceptable at the highest measured noise level of 115 dB and may constitute a problem for other military aviators. The intelligibility degradation due to the noise can be neutralized by use of an available, improved noise-canceling microphone, by the application of current active noise reduction technology to the personal communication equipment, and by the development of a voice communications system to accommodate the speech produced by both female and male aviators.

  7. The value of visualizing tone of voice.

    PubMed

    Pullin, Graham; Cook, Andrew

    2013-10-01

    Whilst most of us have an innate feeling for tone of voice, it is an elusive quality that even phoneticians struggle to describe with sufficient subtlety. For people who cannot speak themselves this can have particularly profound repercussions. Augmentative communication often involves text-to-speech, a technology that only supports a basic choice of prosody based on punctuation. Given how inherently difficult it is to talk about more nuanced tone of voice, there is a risk that its absence from current devices goes unremarked and unchallenged. Looking ahead optimistically to more expressive communication aids, their design will need to involve more subtle interactions with tone of voice-interactions that the people using them can understand and engage with. Interaction design can play a role in making tone of voice visible, tangible, and accessible. Two projects that have already catalysed interdisciplinary debate in this area, Six Speaking Chairs and Speech Hedge, are introduced together with responses. A broader role for design is advocated, as a means to opening up speech technology research to a wider range of disciplinary perspectives, and also to the contributions and influence of people who use it in their everyday lives.

  8. Unique voices in harmony: Call-and-response to address race and physics teaching

    NASA Astrophysics Data System (ADS)

    Cochran, Geraldine L.; White, Gary D.

    2017-09-01

    In the February 2016 issue of The Physics Teacher, we announced a call for papers on race and physics teaching. The response was muted at first, but has now grown to a respectable chorale-sized volume. As the manuscripts began to come in and the review process progressed, Geraldine Cochran graciously agreed to come on board as co-editor for this remarkable collection of papers, to be published throughout the fall of 2017 in TPT. Upon reviewing the original call and the responses from the physics community, the parallels between generating this collection and the grand call-and-response tradition became compelling. What follows is a conversation constructed by the co-editors that is intended to introduce the reader to the swell of voices that responded to the original call. The authors would like to thank Pam Aycock for providing many useful contributions to this editorial.

  9. Dissociating Long and Short-term Memory in Three-Month-Old Infants Using the Mismatch Response to Voice Stimuli

    PubMed Central

    Zinke, Katharina; Thöne, Leonie; Bolinger, Elaina M.; Born, Jan

    2018-01-01

    Auditory event-related potentials (ERPs) have been successfully used in adults as well as in newborns to discriminate recall of longer-term and shorter-term memories. Specifically the Mismatch Response (MMR) to deviant stimuli of an oddball paradigm is larger if the deviant stimuli are highly familiar (i.e., retrieved from long-term memory) than if they are unfamiliar, representing an immediate change to the standard stimuli kept in short-term memory. Here, we aimed to extend previous findings indicating a differential MMR to familiar and unfamiliar deviants in newborns (Beauchemin et al., 2011), to 3-month-old infants who are starting to interact more with their social surroundings supposedly based on forming more (social) long-term representations. Using a voice discrimination paradigm, each infant was repeatedly presented with the word “baby” (400 ms, interstimulus interval: 600 ms, 10 min overall duration) pronounced by three different female speakers. One voice that was unfamiliar to the infants served as the frequently presented “standard” stimulus, whereas another unfamiliar voice served as the “unfamiliar deviant” stimulus, and the voice of the infant’s mother served as the “familiar deviant.” Data collection was successful for 31 infants (mean age = 100 days). The MMR was determined by the difference between the ERP to standard stimuli and the ERP to the unfamiliar and familiar deviant, respectively. The MMR to the familiar deviant (mother’s voice) was larger, i.e., more positive, than that to the unfamiliar deviant between 100 and 400 ms post-stimulus over the frontal and central cortex. However, a genuine MMR differentiating, as a positive deflection, between ERPs to familiar deviants and standard stimuli was only found in the 300–400 ms interval. On the other hand, a genuine MMR differentiating, as a negative deflection, between ERPs to unfamiliar deviants from ERPs to standard stimuli was revealed for the 200–300 ms post-stimulus interval. Overall results confirm a differential MMR response to unfamiliar and familiar deviants in 3-month-olds, with the earlier negative MMR to unfamiliar deviants likely reflecting change detection based on comparison processes in short-term memory, and the later positive MMR to familiar deviants reflecting subsequent long-term memory-based processing of stimulus relevance. PMID:29441032

  10. Dissociating Long and Short-term Memory in Three-Month-Old Infants Using the Mismatch Response to Voice Stimuli.

    PubMed

    Zinke, Katharina; Thöne, Leonie; Bolinger, Elaina M; Born, Jan

    2018-01-01

    Auditory event-related potentials (ERPs) have been successfully used in adults as well as in newborns to discriminate recall of longer-term and shorter-term memories. Specifically the Mismatch Response (MMR) to deviant stimuli of an oddball paradigm is larger if the deviant stimuli are highly familiar (i.e., retrieved from long-term memory) than if they are unfamiliar, representing an immediate change to the standard stimuli kept in short-term memory. Here, we aimed to extend previous findings indicating a differential MMR to familiar and unfamiliar deviants in newborns (Beauchemin et al., 2011), to 3-month-old infants who are starting to interact more with their social surroundings supposedly based on forming more (social) long-term representations. Using a voice discrimination paradigm, each infant was repeatedly presented with the word "baby" (400 ms, interstimulus interval: 600 ms, 10 min overall duration) pronounced by three different female speakers. One voice that was unfamiliar to the infants served as the frequently presented "standard" stimulus, whereas another unfamiliar voice served as the "unfamiliar deviant" stimulus, and the voice of the infant's mother served as the "familiar deviant." Data collection was successful for 31 infants (mean age = 100 days). The MMR was determined by the difference between the ERP to standard stimuli and the ERP to the unfamiliar and familiar deviant, respectively. The MMR to the familiar deviant (mother's voice) was larger, i.e., more positive, than that to the unfamiliar deviant between 100 and 400 ms post-stimulus over the frontal and central cortex. However, a genuine MMR differentiating, as a positive deflection, between ERPs to familiar deviants and standard stimuli was only found in the 300-400 ms interval. On the other hand, a genuine MMR differentiating, as a negative deflection, between ERPs to unfamiliar deviants from ERPs to standard stimuli was revealed for the 200-300 ms post-stimulus interval. Overall results confirm a differential MMR response to unfamiliar and familiar deviants in 3-month-olds, with the earlier negative MMR to unfamiliar deviants likely reflecting change detection based on comparison processes in short-term memory, and the later positive MMR to familiar deviants reflecting subsequent long-term memory-based processing of stimulus relevance.

  11. Voice Based City Panic Button System

    NASA Astrophysics Data System (ADS)

    Febriansyah; Zainuddin, Zahir; Bachtiar Nappu, M.

    2018-03-01

    The development of voice activated panic button application aims to design faster early notification of hazardous condition in community to the nearest police by using speech as the detector where the current application still applies touch-combination on screen and use coordination of orders from control center then the early notification still takes longer time. The method used in this research was by using voice recognition as the user voice detection and haversine formula for the comparison of closest distance between the user and the police. This research was equipped with auto sms, which sent notification to the victim’s relatives, that was also integrated with Google Maps application (GMaps) as the map to the victim’s location. The results show that voice registration on the application reaches 100%, incident detection using speech recognition while the application is running is 94.67% in average, and the auto sms to the victim relatives reaches 100%.

  12. A laryngographic and laryngoscopic study of Northern Vietnamese tones.

    PubMed

    Brunelle, Marc; Nguyên, Duy Duong; Nguyên, Khac Hùng

    2010-01-01

    A laryngographic and laryngoscopic study of tone production in Northern Vietnamese, a language whose tones combine both fundamental frequency (f0) modulations and voice qualities (phonation types), was conducted with 5 male and 5 female speakers. Results show that the f0 contours of Northern Vietnamese tones are not only attributable to changes in vocal fold length and tension (partly through changes in larynx height), but that f0 drops are also largely caused by the glottal configurations responsible for the contrastive voice qualities associated with some of the tones. We also find that voice quality contrasts are mostly due to glottal constriction: they occasionally involve additional ventricular fold incursion and epiglottal constriction, but these articulations are usually absent. Copyright © 2010 S. Karger AG, Basel.

  13. Recruitment and Retention Challenges in a Technology-Based Study with Older Adults Discharged from a Geriatric Rehabilitation Unit.

    PubMed

    McCloskey, Rose; Jarrett, Pamela; Stewart, Connie; Keeping-Burke, Lisa

    2015-01-01

    Technology has the potential to offer support to older adults after being discharged from geriatric rehabilitation. This article highlights recruitment and retention challenges in a study examining an interactive voice response telephone system designed to monitor and support older adults and their informal caregivers following discharge from a geriatric rehabilitation unit. A prospective longitudinal study was planned to examine the feasibility of an interactive voice telephone system in facilitating the transition from rehabilitation to home for older adults and their family caregivers. Patient participants were required to make daily calls into the system. Using standardized instruments, data was to be collected at baseline and during home visits. Older adults and their caregivers may not be willing to learn how to use new technology at the time of hospital discharge. Poor recruitment and retention rates prevented analysis of findings. The importance of recruitment and retention in any study should never be underestimated. Target users of any intervention need to be included in both the design of the intervention and the study examining its benefit. Identifying the issues associated with introducing technology with a group of older rehabilitation patients should assist others who are interested in exploring the role of technology in facilitating hospital discharge. © 2014 Association of Rehabilitation Nurses.

  14. Advocacy, support and survivorship in prostate cancer.

    PubMed

    Dunn, J; Casey, C; Sandoe, D; Hyde, M K; Cheron-Sauer, M-C; Lowe, A; Oliffe, J L; Chambers, S K

    2018-03-01

    Across Australia, prostate cancer support groups (PCSG) have emerged to fill a gap in psychosocial care for men and their families. However, an understanding of the triggers and influencers of the PCSG movement is absent. We interviewed 21 SG leaders (19 PC survivors, two partners), of whom six also attended a focus group, about motivations, experiences, past and future challenges in founding and leading PCSGs. Thematic analysis identified four global themes: illness experience; enacting a supportive response; forming a national collective and challenges. Leaders described men's feelings of isolation and neglect by the health system as the impetus for PCSGs to form and give/receive mutual help. Negotiating health care systems was an early challenge. National affiliation enabled leaders to build a united voice in the health system and establish a group identity and collective voice. Affiliation was supported by a symbiotic relationship with tensions between independence, affiliation and governance. Future challenges were group sustainability and inclusiveness. Study findings describe how a grassroots PCSG movement arose consistent with an embodied health movement perspective. Health care organisations who seek to leverage these community resources need to be cognisant of SG values and purpose if they are to negotiate effective partnerships that maximise mutual benefit. © 2017 The Authors. European Journal of Cancer Care Published by John Wiley & Sons Ltd.

  15. Independent Neuronal Representation of Facial and Vocal Identity in the Monkey Hippocampus and Inferotemporal Cortex.

    PubMed

    Sliwa, Julia; Planté, Aurélie; Duhamel, Jean-René; Wirth, Sylvia

    2016-03-01

    Social interactions make up to a large extent the prime material of episodic memories. We therefore asked how social signals are coded by neurons in the hippocampus. Human hippocampus is home to neurons representing familiar individuals in an abstract and invariant manner ( Quian Quiroga et al. 2009). In contradistinction, activity of rat hippocampal cells is only weakly altered by the presence of other rats ( von Heimendahl et al. 2012; Zynyuk et al. 2012). We probed the activity of monkey hippocampal neurons to faces and voices of familiar and unfamiliar individuals (monkeys and humans). Thirty-one percent of neurons recorded without prescreening responded to faces or to voices. Yet responses to faces were more informative about individuals than responses to voices and neuronal responses to facial and vocal identities were not correlated, indicating that in our sample identity information was not conveyed in an invariant manner like in human neurons. Overall, responses displayed by monkey hippocampal neurons were similar to the ones of neurons recorded simultaneously in inferotemporal cortex, whose role in face perception is established. These results demonstrate that the monkey hippocampus participates in the read-out of social information contrary to the rat hippocampus, but possibly lack an explicit conceptual coding of as found in humans. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  16. Emotionally conditioning the target-speech voice enhances recognition of the target speech under "cocktail-party" listening conditions.

    PubMed

    Lu, Lingxi; Bao, Xiaohan; Chen, Jing; Qu, Tianshu; Wu, Xihong; Li, Liang

    2018-05-01

    Under a noisy "cocktail-party" listening condition with multiple people talking, listeners can use various perceptual/cognitive unmasking cues to improve recognition of the target speech against informational speech-on-speech masking. One potential unmasking cue is the emotion expressed in a speech voice, by means of certain acoustical features. However, it was unclear whether emotionally conditioning a target-speech voice that has none of the typical acoustical features of emotions (i.e., an emotionally neutral voice) can be used by listeners for enhancing target-speech recognition under speech-on-speech masking conditions. In this study we examined the recognition of target speech against a two-talker speech masker both before and after the emotionally neutral target voice was paired with a loud female screaming sound that has a marked negative emotional valence. The results showed that recognition of the target speech (especially the first keyword in a target sentence) was significantly improved by emotionally conditioning the target speaker's voice. Moreover, the emotional unmasking effect was independent of the unmasking effect of the perceived spatial separation between the target speech and the masker. Also, (skin conductance) electrodermal responses became stronger after emotional learning when the target speech and masker were perceptually co-located, suggesting an increase of listening efforts when the target speech was informationally masked. These results indicate that emotionally conditioning the target speaker's voice does not change the acoustical parameters of the target-speech stimuli, but the emotionally conditioned vocal features can be used as cues for unmasking target speech.

  17. Communication-related affective, behavioral, and cognitive reactions in speakers with spasmodic dysphonia.

    PubMed

    Watts, Christopher R; Vanryckeghem, Martine

    2017-12-01

    To investigate the self-perceived affective, behavioral, and cognitive reactions associated with communication of speakers with spasmodic dysphonia as a function of employment status. Prospective cross-sectional investigation. 148 Participants with spasmodic dysphonia (SD) completed an adapted version of the Behavior Assessment Battery (BAB-Voice), a multidimensional assessment of self-perceived reactions to communication. The BAB-Voice consisted of four subtests: the Speech Situation Checklist for A) Emotional Reaction (SSC-ER) and B) Speech Disruption (SSC-SD), C) the Behavior Checklist (BCL), and D) the Communication Attitude Test for Adults (BigCAT). Participants were assigned to groups based on employment status (working versus retired). Descriptive comparison of the BAB-Voice in speakers with SD to previously published non-dysphonic speaker data revealed substantially higher scores associated with SD across all four subtests. Multivariate Analysis of Variance (MANOVA) revealed no significantly different BAB-Voice subtest scores as a function of SD group status (working vs. retired). BAB-Voice scores revealed that speakers with SD experienced substantial impact of their voice disorder on communication attitude, coping behaviors, and affective reactions in speaking situations as reflected in their high BAB scores. These impacts do not appear to be influenced by work status, as speakers with SD who were employed or retired experienced similar levels of affective and behavioral reactions in various speaking situations and cognitive responses. These findings are consistent with previously published pilot data. The specificity of items assessed by means of the BAB-Voice may inform the clinician of valid patient-centered treatment goals which target the impairment extended beyond the physiological dimension. 2b.

  18. The interaction of criminal procedure and outcome.

    PubMed

    Laxminarayan, Malini; Pemberton, Antony

    2014-01-01

    Procedural quality is an important aspect of crime victims' experiences in criminal proceedings and consists of different dimensions. Two of these dimensions are procedural justice (voice) and interpersonal justice (respectful treatment). Social psychological research has suggested that both voice and respectful treatment are moderated by the impact of outcomes of justice procedures on individuals' reactions. To add to this research, we extend this assertion to the criminal justice context, examining the interaction between the assessment of procedural quality and outcome favorability with victim's trust in the legal system and self-esteem. Hierarchical regression analyses reveal that voice, respectful treatment and outcome favorability are predictive of trust in the legal system and self-esteem. Further investigation reveals that being treated with respect is only related to trust in the legal system when outcome favorability is high. Copyright © 2014 Elsevier Ltd. All rights reserved.

  19. Comparing the experience of voices in borderline personality disorder with the experience of voices in a psychotic disorder: A systematic review.

    PubMed

    Merrett, Zalie; Rossell, Susan L; Castle, David J

    2016-07-01

    In clinical settings, there is substantial evidence both clinically and empirically to suggest that approximately 50% of individuals with borderline personality disorder experience auditory verbal hallucinations. However, there is limited research investigating the phenomenology of these voices. The aim of this study was to review and compare our current understanding of auditory verbal hallucinations in borderline personality disorder with auditory verbal hallucinations in patients with a psychotic disorder, to critically analyse existing studies investigating auditory verbal hallucinations in borderline personality disorder and to identify gaps in current knowledge, which will help direct future research. The literature was searched using the electronic database Scopus, PubMed and MEDLINE. Relevant studies were included if they were written in English, were empirical studies specifically addressing auditory verbal hallucinations and borderline personality disorder, were peer reviewed, used only adult humans and sample comprising borderline personality disorder as the primary diagnosis, and included a comparison group with a primary psychotic disorder such as schizophrenia. Our search strategy revealed a total of 16 articles investigating the phenomenology of auditory verbal hallucinations in borderline personality disorder. Some studies provided evidence to suggest that the voice experiences in borderline personality disorder are similar to those experienced by people with schizophrenia, for example, occur inside the head, and often involved persecutory voices. Other studies revealed some differences between schizophrenia and borderline personality disorder voice experiences, with the borderline personality disorder voices sounding more derogatory and self-critical in nature and the voice-hearers' response to the voices were more emotionally resistive. Furthermore, in one study, the schizophrenia group's voices resulted in more disruption in daily functioning. These studies are, however, limited in number and do not provide definitive evidence of these differences. The limited research examining auditory verbal hallucinations experiences in borderline personality disorder poses a significant diagnostic and treatment challenge. A deeper understanding of the precise phenomenological characteristics will help us in terms of diagnostic distinction as well as inform treatments. © The Royal Australian and New Zealand College of Psychiatrists 2016.

  20. Comparing the demands of destination entry using Google Glass and the Samsung Galaxy S4 during simulated driving.

    PubMed

    Beckers, Niek; Schreiner, Sam; Bertrand, Pierre; Mehler, Bruce; Reimer, Bryan

    2017-01-01

    The relative impact of using a Google Glass based voice interface to enter a destination address compared to voice and touch-entry methods using a handheld Samsung Galaxy S4 smartphone was assessed in a driving simulator. Voice entry (Google Glass and Samsung) had lower subjective workload ratings, lower standard deviation of lateral lane position, shorter task durations, faster remote Detection Response Task (DRT) reaction times, lower DRT miss rates, and resulted in less time glancing off-road than the primary visual-manual interaction with the Samsung Touch interface. Comparing voice entry methods, using Google Glass took less time, while glance metrics and reaction time to DRT events responded to were similar. In contrast, DRT miss rate was higher for Google Glass, suggesting that drivers may be under increased distraction levels but for a shorter period of time; whether one or the other equates to an overall safer driving experience is an open question. Copyright © 2016 Elsevier Ltd. All rights reserved.

  1. African Security Challenges: Now and Over the Horizon - Voices from the NGO Community

    DTIC Science & Technology

    2010-11-01

    economies will have a tendency to shift back to dependence on unskilled sectors such as mining, black -market or informal trading and international aid...PEPFAR Watch. Pepfarwatch.org. Rice , A. 2007. “An African Solution.” Nation, June 11. Voices from the NGO Community - 5.19 - African Security...oversight role in this area by parliaments, regardless of formal mandates, roles and responsibilities. In order to avoid the risk of cosmetic changes

  2. Voice: challenging the stigma of addiction; a nursing perspective.

    PubMed

    Paivinen, Helena; Bade, Sherrie

    2008-06-01

    Voice is a collection of art, poetry and narratives created by women living with a history of substance use and addiction. The intent of this collection is to explore women's understanding of harm reduction, to challenge the effects of stigmatization and to explore the experiences of those who have historically been silenced or devalued. Voice was conceived by a group of Kamloops nurses who came together and used their knowledge of mainstream systems, aesthetic knowing, feminism and substance use to guide the development and implementation of this project. During weekly gatherings, women with histories of substance use and addiction worked alongside a nurse in the co-creation of artistic expressions. Gender sensitivity, trust, equality and respect were vital to the success of this process. A selection of the women's art was presented at several venues, including an International Conference on Drug Related Harm, a Nursing Conference and a local art gallery. The positive community response to the women's work contributed to feelings of great pride and enhanced the women's confidence in their ability to express themselves. Throughout this process, women had the opportunity to develop social networks and to become aware of the value that their creative knowledge has to the community in which they live. Gender sensitive programming that is inclusive, participative and promotes women's health is required to fully understand women's experience of substance use and addiction in relation to harm reduction. Participation in projects such as Voice supports and encourages women to make sense of the world they live in and encourages health-promoting activities. The promising outcomes of this project might well be developed by nurses in other settings to further promote the health of women who have traditionally been stigmatized.

  3. [Vocal capabilities of nonprofessional singers evaluated by measurement and superimposition of their speaking, shouting and singing voice range profiles].

    PubMed

    Hacki, T

    1999-09-01

    Voice range profile(VRP) measurement (Phonetography) was used for the evaluation of the vocal capabilities of 41 female (F) and 50 male (M) members of a nonprofessional choir. By means of an automatic VPR measurement system F0 and SPL dB(A) were determined and displayed real time, two-dimensionally. The speaking voice (reading a standard passage as well as counting from the softest to the loudest intensity), the shouting voice (3-4 times shouting a standard sentence) and the singing voice (sustained phonation / la:/ at minimum and maximum intensity level) were measured. The VRPs of these voice modalities were superimposed on the screen and the plot. The averaged values for the speaking VRP: intensity range (F): 48 dB (range 46 soft to 94 dB loud phonation), (M): 52 dB (range 46-98). Pitch range (F): 15 semitones (ST) (Cis3, 138-E4, 329 Hz), (M): 19 ST (E2, 82 Hz-H3, 246 Hz). The average slope for the speaking voice (F): 0,31 ST/dB, (M): 0,36 ST/dB. Shouting VRP highest intensity (F): 106,5 dB, (M): 108,5 dB, highest pitch (F): between Ais4, 466 and H4, 493 Hz. (M): E4, 329 Hz. Average slope for speaking and shouting voice (F): 0,36 ST/dB, (M): 0,39 ST/dB. Singing VRP pitch range (F): 34,6 ST, (M): 37 ST, intensity range (F): 60 dB, (M): 58 dB. The pitch extension of the speaking VRP ranges from 2,9 to 46,2%, speaking and shouting VRPs together with 2,9 to 65% of the pitch range of the singing VRP (F), (M) 2,7-54% and 2,7-67,5% accordingly. The average values for nonprofessional singers reflect an effective but not special use of the phonatory system for the speaking, shouting and singing voice functions with respect to pitch and intensity.

  4. [Effects of a voice metronome on compression rate and depth in telephone assisted, bystander cardiopulmonary resuscitation: an investigator-blinded, 3-armed, randomized, simulation trial].

    PubMed

    van Tulder, Raphael; Roth, Dominik; Krammel, Mario; Laggner, Roberta; Schriefl, Christoph; Kienbacher, Calvin; Lorenzo Hartmann, Alexander; Novosad, Heinz; Constantin Chwojka, Christof; Havel, Christoph; Schreiber, Wolfgang; Herkner, Harald

    2015-01-01

    We investigated the effect on compression rate and depth of a conventional metronome and a voice metronome in simulated telephone-assisted, protocol-driven bystander Cardiopulmonary resucitation (CPR) compared to standard instruction. Thirty-six lay volunteers performed 10 minutes of compression-only CPR in a prospective, investigator-blinded, 3-arm study on a manikin. Participants were randomized either to standard instruction ("push down firmly, 5 cm"), a regular metronome pacing 110 beats per minute (bpm), or a voice metronome continuously prompting "deep-deepdeep- deeper" at 110 bpm. The primary outcome was deviation from the ideal chest compression target range (50 mm compression depth x 100 compressions per minute x 10 minutes = 50 m). Secondary outcomes were CPR quality measures (compression and leaning depth, rate, no-flow times) and participants' related physiological response (heart rate, blood pressure and nine hole peg test and borg scales score). We used a linear regression model to calculate effects. The mean (SD) deviation from the ideal target range (50 m) was -11 (9) m in the standard group, -20 (11) m in the conventional metronome group (adjusted difference [95%, CI], 9.0 [1.2-17.5 m], P=.03), and -18 (9) m in the voice metronome group (adjusted difference, 7.2 [-0.9-15.3] m, P=.08). Secondary outcomes (CPR quality measures and physiological response of participants to CPR performance) showed no significant differences. Compared to standard instruction, the conventional metronome showed a significant negative effect on the chest compression target range. The voice metronome showed a non-significant negative effect and therefore cannot be recommended for regular use in telephone-assisted CPR.

  5. Vocal Qualities in Music Theater Voice: Perceptions of Expert Pedagogues.

    PubMed

    Bourne, Tracy; Kenny, Dianna

    2016-01-01

    To gather qualitative descriptions of music theater vocal qualities including belt, legit, and mix from expert pedagogues to better define this voice type. This is a prospective, semistructured interview. Twelve expert teachers from United States, United Kingdom, Asia, and Australia were interviewed by Skype and asked to identify characteristics of music theater vocal qualities including vocal production, physiology, esthetics, pitch range, and pedagogical techniques. Responses were compared with published studies on music theater voice. Belt and legit were generally described as distinct sounds with differing physiological and technical requirements. Teachers were concerned that belt should be taught "safely" to minimize vocal health risks. There was consensus between teachers and published research on the physiology of the glottis and vocal tract; however, teachers were not in agreement about breathing techniques. Neither were teachers in agreement about the meaning of "mix." Most participants described belt as heavily weighted, thick folds, thyroarytenoid-dominant, or chest register; however, there was no consensus on an appropriate term. Belt substyles were named and generally categorized by weightedness or tone color. Descriptions of male belt were less clear than for female belt. This survey provides an overview of expert pedagogical perspectives on the characteristics of belt, legit, and mix qualities in the music theater voice. Although teacher responses are generally in agreement with published research, there are still many controversial issues and gaps in knowledge and understanding of this vocal technique. Breathing techniques, vocal range, mix, male belt, and vocal registers require continuing investigation so that we can learn more about efficient and healthy vocal function in music theater singing. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  6. Using Rate of Divergence as an Objective Measure to Differentiate between Voice Signal Types Based on the Amount of Disorder in the Signal.

    PubMed

    Calawerts, William M; Lin, Liyu; Sprott, J C; Jiang, Jack J

    2017-01-01

    The purpose of this paper is to introduce the rate of divergence as an objective measure to differentiate between the four voice types based on the amount of disorder present in a signal. We hypothesized that rate of divergence would provide an objective measure that can quantify all four voice types. A total of 150 acoustic voice recordings were randomly selected and analyzed using traditional perturbation, nonlinear, and rate of divergence analysis methods. We developed a new parameter, rate of divergence, which uses a modified version of Wolf's algorithm for calculating Lyapunov exponents of a system. The outcome of this calculation is not a Lyapunov exponent, but rather a description of the divergence of two nearby data points for the next three points in the time series, followed in three time-delayed embedding dimensions. This measure was compared to currently existing perturbation and nonlinear dynamic methods of distinguishing between voice signals. There was a direct relationship between voice type and rate of divergence. This calculation is especially effective at differentiating between type 3 and type 4 voices (P < 0.001) and is equally effective at differentiating type 1, type 2, and type 3 signals as currently existing methods. The rate of divergence calculation introduced is an objective measure that can be used to distinguish between all four voice types based on the amount of disorder present, leading to quicker and more accurate voice typing as well as an improved understanding of the nonlinear dynamics involved in phonation. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

  7. A voice coil actuator driven active vibration isolation system with the consideration of flexible modes.

    PubMed

    Park, Kyihwan; Choi, Dongyoub; Ozer, Abdullah; Kim, Sangyoo; Lee, Yongkwan; Joo, Dongik

    2008-06-01

    We develop a four-mount active vibration isolation system (AVIS) using voice coil actuators. The flexible body modes in the upper plate of the AVIS can cause an instability problem due to control signal whose frequency is close to the resonant frequency of the flexible modes. The loop shaping technique is applied to reduce the amplitude of the control signal. We investigate the performances of the active vibration isolation system proposed in the word in the time domain and frequency domain by comparing to the passive isolation system.

  8. System Design Plan for a DCS (Defense Communications System) Data Transmission Network.

    DTIC Science & Technology

    1981-07-01

    modems , FDO group modems , and Voice Frequency Carrier Telegraph (VFCT) networks. The DTN will be a synchronous network and its implementation must coincide...Frequency (VF) modems and Voice Frequency Carrier Telegraph (VFCT) networks. Further, data circuits can be extended over present analog FDM facilities using...VF or group data modems . In addition to the availability of terrestrial and satellite digital transmission facilities, the implementation of the DTN

  9. Computer-Based Instruction for TRIDENT FBM Training

    DTIC Science & Technology

    1976-06-01

    remote voice feedback to an operator. In this case it is possible to display text which represents the voice messages required during sequential ...provides two main services: (a) the preparation of missiles for sequential launching with self-guidance after launch, and (b) the coordination of...monitor- ing the status of the guidance system in each missile. FCS SWS coordina- tion consists of monitoring systems involved in sequential functions at

  10. High precision locating control system based on VCM for Talbot lithography

    NASA Astrophysics Data System (ADS)

    Yao, Jingwei; Zhao, Lixin; Deng, Qian; Hu, Song

    2016-10-01

    Aiming at the high precision and efficiency requirements of Z-direction locating in Talbot lithography, a control system based on Voice Coil Motor (VCM) was designed. In this paper, we built a math model of VCM and its moving characteristic was analyzed. A double-closed loop control strategy including position loop and current loop were accomplished. The current loop was implemented by driver, in order to achieve the rapid follow of the system current. The position loop was completed by the digital signal processor (DSP) and the position feedback was achieved by high precision linear scales. Feed forward control and position feedback Proportion Integration Differentiation (PID) control were applied in order to compensate for dynamic lag and improve the response speed of the system. And the high precision and efficiency of the system were verified by simulation and experiments. The results demonstrated that the performance of Z-direction gantry was obviously improved, having high precision, quick responses, strong real-time and easily to expend for higher precision.

  11. Student Telephone Self-Activation at Boston College.

    ERIC Educational Resources Information Center

    McCormack, David

    1997-01-01

    By viewing individual communication services (voice, data, cable) as strategic tools in a communication infrastructure, Boston College (Massachusetts) was able to create an electronic communication environment with superior services at drastically reduced cost. The system provides voice, data, and cable access to every residence hall room,…

  12. Voice Enabled Framework to Support Post-Surgical Discharge Monitoring

    PubMed Central

    Blansit, Kevin; Marmor, Rebecca; Zhao, Beiqun; Tien, Dan

    2017-01-01

    Unplanned surgical readmissions pose a challenging problem for the American healthcare system. We propose to combine consumer electronic voice recognition technology with the FHIR standard to create a post-surgical discharge monitoring app to identify and alert physicians to a patient’s deteriorating status. PMID:29854267

  13. Central voice production and pathophysiology of spasmodic dysphonia.

    PubMed

    Mor, Niv; Simonyan, Kristina; Blitzer, Andrew

    2018-01-01

    Our ability to speak is complex, and the role of the central nervous system in controlling speech production is often overlooked in the field of otolaryngology. In this brief review, we present an integrated overview of speech production with a focus on the role of central nervous system. The role of central control of voice production is then further discussed in relation to the potential pathophysiology of spasmodic dysphonia (SD). Peer-review articles on central laryngeal control and SD were identified from PUBMED search. Selected articles were augmented with designated relevant publications. Publications that discussed central and peripheral nervous system control of voice production and the central pathophysiology of laryngeal dystonia were chosen. Our ability to speak is regulated by specialized complex mechanisms coordinated by high-level cortical signaling, brainstem reflexes, peripheral nerves, muscles, and mucosal actions. Recent studies suggest that SD results from a primary central disturbance associated with dysfunction at our highest levels of central voice control. The efficacy of botulinum toxin in treating SD may not be limited solely to its local effect on laryngeal muscles and also may modulate the disorder at the level of the central nervous system. Future therapeutic options that target the central nervous system may help modulate the underlying disorder in SD and allow clinicians to better understand the principal pathophysiology. NA.Laryngoscope, 128:177-183, 2018. © 2017 The American Laryngological, Rhinological and Otological Society, Inc.

  14. Interference effects of vocalization on dual task performance

    NASA Astrophysics Data System (ADS)

    Owens, J. M.; Goodman, L. S.; Pianka, M. J.

    1984-09-01

    Voice command and control systems have been proposed as a potential means of off-loading the typically overburdened visual information processing system. However, prior to introducing novel human-machine interfacing technologies in high workload environments, consideration must be given to the integration of the new technologists within existing task structures to ensure that no new sources of workload or interference are systematically introduced. This study examined the use of voice interactive systems technology in the joint performance of two cognitive information processing tasks requiring continuous memory and choice reaction wherein a basis for intertask interference might be expected. Stimuli for the continuous memory task were presented aurally and either voice or keyboard responding was required in the choice reaction task. Performance was significantly degraded in each task when voice responding was required in the choice reaction time task. Performance degradation was evident in higher error scores for both the choice reaction and continuous memory tasks. Performance decrements observed under conditions of high intertask stimulus similarity were not statistically significant. The results signal the need to consider further the task requirements for verbal short-term memory when applying speech technology in multitask environments.

  15. Systemic Hydration: Relating Science to Clinical Practice in Vocal Health

    PubMed Central

    Hartley, Naomi A.; Thibeault, Susan L.

    2014-01-01

    Objectives To examine the current state of the science regarding the role of systemic hydration in vocal function and health. Study Design Literature Review Methods Literature search spanning multiple disciplines, including speech-language pathology, nutrition and dietetics, medicine, sports and exercise science, physiology and biomechanics. Results The relationship between hydration and physical function is an area of common interest amongst multiple professions. Each discipline provides valuable insight into the connection between performance and water balance, as well as complimentary methods of investigation. Existing voice literature suggests a relationship between hydration and voice production, however the underlying mechanisms are not yet defined and a treatment effect for systemic hydration remains to be demonstrated. Literature from other disciplines sheds light on methodological shortcomings and in some cases offers an alternative explanation for observed phenomena. Conclusions A growing body of literature in the field of voice science is documenting a relationship between hydration and vocal function, however greater understanding is required to guide best practice in the maintenance of vocal health and management of voice disorders. Integration of knowledge and technical expertise from multiple disciplines facilitates analysis of existing literature and provides guidance as to future research. PMID:24880674

  16. FonaDyn - A system for real-time analysis of the electroglottogram, over the voice range

    NASA Astrophysics Data System (ADS)

    Ternström, Sten; Johansson, Dennis; Selamtzis, Andreas

    2018-01-01

    From soft to loud and low to high, the mechanisms of human voice have many degrees of freedom, making it difficult to assess phonation from the acoustic signal alone. FonaDyn is a research tool that combines acoustics with electroglottography (EGG). It characterizes and visualizes in real time the dynamics of EGG waveforms, using statistical clustering of the cycle-synchronous EGG Fourier components, and their sample entropy. The prevalence and stability of different EGG waveshapes are mapped as colored regions into a so-called voice range profile, without needing pre-defined thresholds or categories. With appropriately 'trained' clusters, FonaDyn can classify and map voice regimes. This is of potential scientific, clinical and pedagogical interest.

  17. Discourse-voice regulatory strategies in the psychotherapeutic interaction: a state-space dynamics analysis.

    PubMed

    Tomicic, Alemka; Martínez, Claudio; Pérez, J Carola; Hollenstein, Tom; Angulo, Salvador; Gerstmann, Adam; Barroux, Isabelle; Krause, Mariane

    2015-01-01

    This study seeks to provide evidence of the dynamics associated with the configurations of discourse-voice regulatory strategies in patient-therapist interactions in relevant episodes within psychotherapeutic sessions. Its central assumption is that discourses manifest themselves differently in terms of their prosodic characteristics according to their regulatory functions in a system of interactions. The association between discourse and vocal quality in patients and therapists was analyzed in a sample of 153 relevant episodes taken from 164 sessions of five psychotherapies using the state space grid (SSG) method, a graphical tool based on the dynamic systems theory (DST). The results showed eight recurrent and stable discourse-voice regulatory strategies of the patients and three of the therapists. Also, four specific groups of these discourse-voice strategies were identified. The latter were interpreted as regulatory configurations, that is to say, as emergent self-organized groups of discourse-voice regulatory strategies constituting specific interactional systems. Both regulatory strategies and their configurations differed between two types of relevant episodes: Change Episodes and Rupture Episodes. As a whole, these results support the assumption that speaking and listening, as dimensions of the interaction that takes place during therapeutic conversation, occur at different levels. The study not only shows that these dimensions are dependent on each other, but also that they function as a complex and dynamic whole in therapeutic dialog, generating relational offers which allow the patient and the therapist to regulate each other and shape the psychotherapeutic process that characterizes each type of relevant episode.

  18. Using rate of divergence as an objective measure to differentiate between voice signal types based on the amount of disorder in the signal

    PubMed Central

    Calawerts, William M; Lin, Liyu; Sprott, JC; Jiang, Jack J

    2016-01-01

    Objective/Hypothesis The purpose of this paper is to introduce rate of divergence as an objective measure to differentiate between the four voice types based on the amount of disorder present in a signal. We hypothesized that rate of divergence would provide an objective measure that can quantify all four voice types. Study Design 150 acoustic voice recordings were randomly selected and analyzed using traditional perturbation, nonlinear, and rate of divergence analysis methods. ty Methods We developed a new parameter, rate of divergence, which uses a modified version of Wolf’s algorithm for calculating Lyapunov exponents of a system. The outcome of this calculation is not a Lyapunov exponent, but rather a description of the divergence of two nearby data points for the next three points in the time series, followed in three time delayed embedding dimensions. This measure was compared to currently existing perturbation and nonlinear dynamic methods of distinguishing between voice signals. Results There was a direct relationship between voice type and rate of divergence. This calculation is especially effective at differentiating between type 3 and type 4 voices (p<0.001), and is equally effective at differentiating type 1, type 2, and type 3 signals as currently existing methods. Conclusion The rate of divergence calculation introduced is an objective measure that can be used to distinguish between all four voice types based on amount of disorder present, leading to quicker and more accurate voice typing as well as an improved understanding of the nonlinear dynamics involved in phonation. PMID:26920858

  19. Association of trait emotional intelligence and individual fMRI-activation patterns during the perception of social signals from voice and face.

    PubMed

    Kreifelts, Benjamin; Ethofer, Thomas; Huberle, Elisabeth; Grodd, Wolfgang; Wildgruber, Dirk

    2010-07-01

    Multimodal integration of nonverbal social signals is essential for successful social interaction. Previous studies have implicated the posterior superior temporal sulcus (pSTS) in the perception of social signals such as nonverbal emotional signals as well as in social cognitive functions like mentalizing/theory of mind. In the present study, we evaluated the relationships between trait emotional intelligence (EI) and fMRI activation patterns in individual subjects during the multimodal perception of nonverbal emotional signals from voice and face. Trait EI was linked to hemodynamic responses in the right pSTS, an area which also exhibits a distinct sensitivity to human voices and faces. Within all other regions known to subserve the perceptual audiovisual integration of human social signals (i.e., amygdala, fusiform gyrus, thalamus), no such linked responses were observed. This functional difference in the network for the audiovisual perception of human social signals indicates a specific contribution of the pSTS as a possible interface between the perception of social information and social cognition. (c) 2009 Wiley-Liss, Inc.

  20. Analysis of wolves and sheep. Final report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hogden, J.; Papcun, G.; Zlokarnik, I.

    1997-08-01

    In evaluating speaker verification systems, asymmetries have been observed in the ease with which people are able to break into other people`s voice locks. People who are good at breaking into voice locks are called wolves, and people whose locks are easy to break into are called sheep. (Goats are people that have a difficult time opening their own voice locks.) Analyses of speaker verification algorithms could be used to understand wolf/sheep asymmetries. Using the notion of a ``speaker space``, it is demonstrated that such asymmetries could arise even though the similarity of voice 1 to voice 2 is themore » same as the inverse similarity. This explains partially the wolf/sheep asymmetries, although there may be other factors. The speaker space can be computed from interspeaker similarity data using multidimensional scaling, and such speaker space can be used to given a good approximation of the interspeaker similarities. The derived speaker space can be used to predict which of the enrolled speakers are likely to be wolves and which are likely to be sheep. However, a speaker must first enroll in the speaker key system and then be compared to each of the other speakers; a good estimate of a person`s speaker space position could be obtained using only a speech sample.« less

Top