Science.gov

Sample records for acoustics speech communication

  1. Analog Acoustic Expression in Speech Communication

    ERIC Educational Resources Information Center

    Shintel, Hadas; Nusbaum, Howard C.; Okrent, Arika

    2006-01-01

    We present the first experimental evidence of a phenomenon in speech communication we call "analog acoustic expression." Speech is generally thought of as conveying information in two distinct ways: discrete linguistic-symbolic units such as words and sentences represent linguistic meaning, and continuous prosodic forms convey information about…

  2. Acoustical conditions for speech communication in active elementary school classrooms

    NASA Astrophysics Data System (ADS)

    Sato, Hiroshi; Bradley, John

    2005-04-01

    Detailed acoustical measurements were made in 34 active elementary school classrooms with typical rectangular room shape in schools near Ottawa, Canada. There was an average of 21 students in classrooms. The measurements were made to obtain accurate indications of the acoustical quality of conditions for speech communication during actual teaching activities. Mean speech and noise levels were determined from the distribution of recorded sound levels and the average speech-to-noise ratio was 11 dBA. Measured mid-frequency reverberation times (RT) during the same occupied conditions varied from 0.3 to 0.6 s, and were a little less than for the unoccupied rooms. RT values were not related to noise levels. Octave band speech and noise levels, useful-to-detrimental ratios, and Speech Transmission Index values were also determined. Key results included: (1) The average vocal effort of teachers corresponded to louder than Pearsons Raised voice level; (2) teachers increase their voice level to overcome ambient noise; (3) effective speech levels can be enhanced by up to 5 dB by early reflection energy; and (4) student activity is seen to be the dominant noise source, increasing average noise levels by up to 10 dBA during teaching activities. [Work supported by CLLRnet.

  3. Evaluation of acoustical conditions for speech communication in working elementary school classrooms.

    PubMed

    Sato, Hiroshi; Bradley, John S

    2008-04-01

    Detailed acoustical measurements were made in 41 working elementary school classrooms near Ottawa, Canada to obtain more representative and more accurate indications of the acoustical quality of conditions for speech communication during actual teaching activities. This paper describes the room acoustics characteristics and noise environment of 27 traditional rectangular classrooms from the 41 measured rooms. The purpose of the work was to better understand how to improve speech communication between teachers and students. The study found, that on average, the students experienced: teacher speech levels of 60.4 dB A, noise levels of 49.1 dB A, and a mean speech-to-noise ratio of 11 dB A during teaching activities. The mean reverberation time in the occupied classrooms was 0.41 s, which was 10% less than in the unoccupied rooms. The reverberation time measurements were used to determine the average absorption added by each student. Detailed analyses of early and late-arriving speech sounds showed these sound levels could be predicted quite accurately and suggest improved approaches to room acoustics design.

  4. Disordered speech disrupts conversational entrainment: a study of acoustic-prosodic entrainment and communicative success in populations with communication challenges.

    PubMed

    Borrie, Stephanie A; Lubold, Nichola; Pon-Barry, Heather

    2015-01-01

    Conversational entrainment, a pervasive communication phenomenon in which dialogue partners adapt their behaviors to align more closely with one another, is considered essential for successful spoken interaction. While well-established in other disciplines, this phenomenon has received limited attention in the field of speech pathology and the study of communication breakdowns in clinical populations. The current study examined acoustic-prosodic entrainment, as well as a measure of communicative success, in three distinctly different dialogue groups: (i) healthy native vs. healthy native speakers (Control), (ii) healthy native vs. foreign-accented speakers (Accented), and (iii) healthy native vs. dysarthric speakers (Disordered). Dialogue group comparisons revealed significant differences in how the groups entrain on particular acoustic-prosodic features, including pitch, intensity, and jitter. Most notably, the Disordered dialogues were characterized by significantly less acoustic-prosodic entrainment than the Control dialogues. Further, a positive relationship between entrainment indices and communicative success was identified. These results suggest that the study of conversational entrainment in speech pathology will have essential implications for both scientific theory and clinical application in this domain.

  5. Disordered speech disrupts conversational entrainment: a study of acoustic-prosodic entrainment and communicative success in populations with communication challenges

    PubMed Central

    Borrie, Stephanie A.; Lubold, Nichola; Pon-Barry, Heather

    2015-01-01

    Conversational entrainment, a pervasive communication phenomenon in which dialogue partners adapt their behaviors to align more closely with one another, is considered essential for successful spoken interaction. While well-established in other disciplines, this phenomenon has received limited attention in the field of speech pathology and the study of communication breakdowns in clinical populations. The current study examined acoustic-prosodic entrainment, as well as a measure of communicative success, in three distinctly different dialogue groups: (i) healthy native vs. healthy native speakers (Control), (ii) healthy native vs. foreign-accented speakers (Accented), and (iii) healthy native vs. dysarthric speakers (Disordered). Dialogue group comparisons revealed significant differences in how the groups entrain on particular acoustic–prosodic features, including pitch, intensity, and jitter. Most notably, the Disordered dialogues were characterized by significantly less acoustic-prosodic entrainment than the Control dialogues. Further, a positive relationship between entrainment indices and communicative success was identified. These results suggest that the study of conversational entrainment in speech pathology will have essential implications for both scientific theory and clinical application in this domain. PMID:26321996

  6. Speech communications in noise

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The physical characteristics of speech, the methods of speech masking measurement, and the effects of noise on speech communication are investigated. Topics include the speech signal and intelligibility, the effects of noise on intelligibility, the articulation index, and various devices for evaluating speech systems.

  7. Careers in Speech Communication.

    ERIC Educational Resources Information Center

    Speech Communication Association, New York, NY.

    Brief discussions in this pamphlet suggest educational and career opportunities in the following fields of speech communication: rhetoric, public address, and communication; theatre, drama, and oral interpretation; radio, television, and film; speech pathology and audiology; speech science, phonetics, and linguistics; and speech education.…

  8. Acoustic characteristics of listener-constrained speech

    NASA Astrophysics Data System (ADS)

    Ashby, Simone; Cummins, Fred

    2003-04-01

    Relatively little is known about the acoustical modifications speakers employ to meet the various constraints-auditory, linguistic and otherwise-of their listeners. Similarly, the manner by which perceived listener constraints interact with speakers' adoption of specialized speech registers is poorly Hypo (H&H) theory offers a framework for examining the relationship between speech production and output-oriented goals for communication, suggesting that under certain circumstances speakers may attempt to minimize phonetic ambiguity by employing a ``hyperarticulated'' speaking style (Lindblom, 1990). It remains unclear, however, what the acoustic correlates of hyperarticulated speech are, and how, if at all, we might expect phonetic properties to change respective to different listener-constrained conditions. This paper is part of a preliminary investigation concerned with comparing the prosodic characteristics of speech produced across a range of listener constraints. Analyses are drawn from a corpus of read hyperarticulated speech data comprising eight adult, female speakers of English. Specialized registers include speech to foreigners, infant-directed speech, speech produced under noisy conditions, and human-machine interaction. The authors gratefully acknowledge financial support of the Irish Higher Education Authority, allocated to Fred Cummins for collaborative work with Media Lab Europe.

  9. Speech and Communication Disorders

    MedlinePlus

    ... or understand speech. Causes include Hearing disorders and deafness Voice problems, such as dysphonia or those caused ... language therapy can help. NIH: National Institute on Deafness and Other Communication Disorders

  10. Measuring Speech Communication Skills.

    ERIC Educational Resources Information Center

    Carpenter, Edwin C.

    Improving the quality of undergraduate speech communication education depends to a large extent on effectively measuring student achievement in college level communication skills. While formal tests are not as well developed for speaking skills as for other areas of the curriculum, they are available. The two used most frequently are the…

  11. Research in speech communication.

    PubMed Central

    Flanagan, J

    1995-01-01

    Advances in digital speech processing are now supporting application and deployment of a variety of speech technologies for human/machine communication. In fact, new businesses are rapidly forming about these technologies. But these capabilities are of little use unless society can afford them. Happily, explosive advances in microelectronics over the past two decades have assured affordable access to this sophistication as well as to the underlying computing technology. The research challenges in speech processing remain in the traditionally identified areas of recognition, synthesis, and coding. These three areas have typically been addressed individually, often with significant isolation among the efforts. But they are all facets of the same fundamental issue--how to represent and quantify the information in the speech signal. This implies deeper understanding of the physics of speech production, the constraints that the conventions of language impose, and the mechanism for information processing in the auditory system. In ongoing research, therefore, we seek more accurate models of speech generation, better computational formulations of language, and realistic perceptual guides for speech processing--along with ways to coalesce the fundamental issues of recognition, synthesis, and coding. Successful solution will yield the long-sought dictation machine, high-quality synthesis from text, and the ultimate in low bit-rate transmission of speech. It will also open the door to language-translating telephony, where the synthetic foreign translation can be in the voice of the originating talker. Images Fig. 1 Fig. 2 Fig. 5 Fig. 8 Fig. 11 Fig. 12 Fig. 13 PMID:7479806

  12. Sperry Univac speech communications technology

    NASA Technical Reports Server (NTRS)

    Medress, Mark F.

    1977-01-01

    Technology and systems for effective verbal communication with computers were developed. A continuous speech recognition system for verbal input, a word spotting system to locate key words in conversational speech, prosodic tools to aid speech analysis, and a prerecorded voice response system for speech output are described.

  13. Acoustic Differences between Humorous and Sincere Communicative Intentions

    ERIC Educational Resources Information Center

    Hoicka, Elena; Gattis, Merideth

    2012-01-01

    Previous studies indicate that the acoustic features of speech discriminate between positive and negative communicative intentions, such as approval and prohibition. Two studies investigated whether acoustic features of speech can discriminate between two positive communicative intentions: humour and sweet-sincerity, where sweet-sincerity involved…

  14. Acoustics of Clear Speech: Effect of Instruction

    ERIC Educational Resources Information Center

    Lam, Jennifer; Tjaden, Kris; Wilding, Greg

    2012-01-01

    Purpose: This study investigated how different instructions for eliciting clear speech affected selected acoustic measures of speech. Method: Twelve speakers were audio-recorded reading 18 different sentences from the Assessment of Intelligibility of Dysarthric Speech (Yorkston & Beukelman, 1984). Sentences were produced in habitual, clear,…

  15. Acoustic communication by ants

    NASA Astrophysics Data System (ADS)

    Hickling, Robert

    2002-05-01

    Many ant species communicate acoustically by stridulating, i.e., running a scraper over a washboard-like set of ridges. Ants appear to be insensitive to airborne sound. Consequently, myrmecologists have concluded that the stridulatory signals are transmitted through the substrate. This has tended to diminish the importance of acoustic communication, and it is currently believed that ant communication is based almost exclusively on pheromones, with acoustic communication assigned an almost nonexistent role. However, it can be shown that acoustic communication between ants is effective only if the medium is air and not the substrate. How, then, is it possible for ants to appear deaf to airborne sound and yet communicate through the air? An explanation is provided in a paper [R. Hickling and R. L. Brown, ``Analysis of acoustic communication by ants,'' J. Acoust. Soc. Am. 108, 1920-1929 (2000)]. Ants are small relative to the wavelengths they generate. Hence, they create a near field, which is characterized by a major increase in sound velocity (particle velocity of sound) in the vicinity of the source. Hair sensilla on the ants' antennae respond to sound velocity. Thus, ants are able to detect near-field sound from other ants and to exclude extraneous airborne sound.

  16. Brain-Computer Interfaces for Speech Communication

    PubMed Central

    Brumberg, Jonathan S.; Nieto-Castanon, Alfonso; Kennedy, Philip R.; Guenther, Frank H.

    2010-01-01

    This paper briefly reviews current silent speech methodologies for normal and disabled individuals. Current techniques utilizing electromyographic (EMG) recordings of vocal tract movements are useful for physically healthy individuals but fail for tetraplegic individuals who do not have accurate voluntary control over the speech articulators. Alternative methods utilizing EMG from other body parts (e.g., hand, arm, or facial muscles) or electroencephalography (EEG) can provide capable silent communication to severely paralyzed users, though current interfaces are extremely slow relative to normal conversation rates and require constant attention to a computer screen that provides visual feedback and/or cueing. We present a novel approach to the problem of silent speech via an intracortical microelectrode brain computer interface (BCI) to predict intended speech information directly from the activity of neurons involved in speech production. The predicted speech is synthesized and acoustically fed back to the user with a delay under 50 ms. We demonstrate that the Neurotrophic Electrode used in the BCI is capable of providing useful neural recordings for over 4 years, a necessary property for BCIs that need to remain viable over the lifespan of the user. Other design considerations include neural decoding techniques based on previous research involving BCIs for computer cursor or robotic arm control via prediction of intended movement kinematics from motor cortical signals in monkeys and humans. Initial results from a study of continuous speech production with instantaneous acoustic feedback show the BCI user was able to improve his control over an artificial speech synthesizer both within and across recording sessions. The success of this initial trial validates the potential of the intracortical microelectrode-based approach for providing a speech prosthesis that can allow much more rapid communication rates. PMID:20204164

  17. Brain-Computer Interfaces for Speech Communication.

    PubMed

    Brumberg, Jonathan S; Nieto-Castanon, Alfonso; Kennedy, Philip R; Guenther, Frank H

    2010-04-01

    This paper briefly reviews current silent speech methodologies for normal and disabled individuals. Current techniques utilizing electromyographic (EMG) recordings of vocal tract movements are useful for physically healthy individuals but fail for tetraplegic individuals who do not have accurate voluntary control over the speech articulators. Alternative methods utilizing EMG from other body parts (e.g., hand, arm, or facial muscles) or electroencephalography (EEG) can provide capable silent communication to severely paralyzed users, though current interfaces are extremely slow relative to normal conversation rates and require constant attention to a computer screen that provides visual feedback and/or cueing. We present a novel approach to the problem of silent speech via an intracortical microelectrode brain computer interface (BCI) to predict intended speech information directly from the activity of neurons involved in speech production. The predicted speech is synthesized and acoustically fed back to the user with a delay under 50 ms. We demonstrate that the Neurotrophic Electrode used in the BCI is capable of providing useful neural recordings for over 4 years, a necessary property for BCIs that need to remain viable over the lifespan of the user. Other design considerations include neural decoding techniques based on previous research involving BCIs for computer cursor or robotic arm control via prediction of intended movement kinematics from motor cortical signals in monkeys and humans. Initial results from a study of continuous speech production with instantaneous acoustic feedback show the BCI user was able to improve his control over an artificial speech synthesizer both within and across recording sessions. The success of this initial trial validates the potential of the intracortical microelectrode-based approach for providing a speech prosthesis that can allow much more rapid communication rates.

  18. Speech coding, reconstruction and recognition using acoustics and electromagnetic waves

    DOEpatents

    Holzrichter, J.F.; Ng, L.C.

    1998-03-17

    The use of EM radiation in conjunction with simultaneously recorded acoustic speech information enables a complete mathematical coding of acoustic speech. The methods include the forming of a feature vector for each pitch period of voiced speech and the forming of feature vectors for each time frame of unvoiced, as well as for combined voiced and unvoiced speech. The methods include how to deconvolve the speech excitation function from the acoustic speech output to describe the transfer function each time frame. The formation of feature vectors defining all acoustic speech units over well defined time frames can be used for purposes of speech coding, speech compression, speaker identification, language-of-speech identification, speech recognition, speech synthesis, speech translation, speech telephony, and speech teaching. 35 figs.

  19. Speech coding, reconstruction and recognition using acoustics and electromagnetic waves

    DOEpatents

    Holzrichter, John F.; Ng, Lawrence C.

    1998-01-01

    The use of EM radiation in conjunction with simultaneously recorded acoustic speech information enables a complete mathematical coding of acoustic speech. The methods include the forming of a feature vector for each pitch period of voiced speech and the forming of feature vectors for each time frame of unvoiced, as well as for combined voiced and unvoiced speech. The methods include how to deconvolve the speech excitation function from the acoustic speech output to describe the transfer function each time frame. The formation of feature vectors defining all acoustic speech units over well defined time frames can be used for purposes of speech coding, speech compression, speaker identification, language-of-speech identification, speech recognition, speech synthesis, speech translation, speech telephony, and speech teaching.

  20. Optimizing acoustical conditions for speech intelligibility in classrooms

    NASA Astrophysics Data System (ADS)

    Yang, Wonyoung

    High speech intelligibility is imperative in classrooms where verbal communication is critical. However, the optimal acoustical conditions to achieve a high degree of speech intelligibility have previously been investigated with inconsistent results, and practical room-acoustical solutions to optimize the acoustical conditions for speech intelligibility have not been developed. This experimental study validated auralization for speech-intelligibility testing, investigated the optimal reverberation for speech intelligibility for both normal and hearing-impaired listeners using more realistic room-acoustical models, and proposed an optimal sound-control design for speech intelligibility based on the findings. The auralization technique was used to perform subjective speech-intelligibility tests. The validation study, comparing auralization results with those of real classroom speech-intelligibility tests, found that if the room to be auralized is not very absorptive or noisy, speech-intelligibility tests using auralization are valid. The speech-intelligibility tests were done in two different auralized sound fields---approximately diffuse and non-diffuse---using the Modified Rhyme Test and both normal and hearing-impaired listeners. A hybrid room-acoustical prediction program was used throughout the work, and it and a 1/8 scale-model classroom were used to evaluate the effects of ceiling barriers and reflectors. For both subject groups, in approximately diffuse sound fields, when the speech source was closer to the listener than the noise source, the optimal reverberation time was zero. When the noise source was closer to the listener than the speech source, the optimal reverberation time was 0.4 s (with another peak at 0.0 s) with relative output power levels of the speech and noise sources SNS = 5 dB, and 0.8 s with SNS = 0 dB. In non-diffuse sound fields, when the noise source was between the speaker and the listener, the optimal reverberation time was 0.6 s with

  1. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2006-08-08

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  2. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2004-03-23

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  3. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2006-02-14

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  4. Speech Communication and Telephone Networks

    NASA Astrophysics Data System (ADS)

    Gierlich, H. W.

    Speech communication over telephone networks has one major constraint: The communication has to be “real time”. The basic principle since the beginning of all telephone networks has been to provide a communication system capable of substituting the air path between two persons having a conversation at 1-m distance. This is the so-called orthotelephonic reference position [7]. Although many technical compromises must be made to enable worldwide communication over telephone networks, it is still the goal to achieve speech quality performance which is close to this reference.

  5. Critical Thinking in Speech Communication: Survey of Speech Communication Educators.

    ERIC Educational Resources Information Center

    Ruminski, Henry; And Others

    The interests of all communication educators would best be served if those educators could agree on a single broad definition of critical thinking that incorporates a variety of perspectives. Toward this end, questionnaires were sent to a ramdom sampling of 300 members of the Speech Communication Association; 88 were returned. The questionnaire…

  6. Speech Intelligibility Advantages using an Acoustic Beamformer Display

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Sunder, Kaushik; Godfroy, Martine; Otto, Peter

    2015-01-01

    A speech intelligibility test conforming to the Modified Rhyme Test of ANSI S3.2 "Method for Measuring the Intelligibility of Speech Over Communication Systems" was conducted using a prototype 12-channel acoustic beamformer system. The target speech material (signal) was identified against speech babble (noise), with calculated signal-noise ratios of 0, 5 and 10 dB. The signal was delivered at a fixed beam orientation of 135 deg (re 90 deg as the frontal direction of the array) and the noise at 135 deg (co-located) and 0 deg (separated). A significant improvement in intelligibility from 57% to 73% was found for spatial separation for the same signal-noise ratio (0 dB). Significant effects for improved intelligibility due to spatial separation were also found for higher signal-noise ratios (5 and 10 dB).

  7. Speech recognition: Acoustic, phonetic and lexical

    NASA Astrophysics Data System (ADS)

    Zue, V. W.

    1985-10-01

    Our long-term research goal is the development and implementation of speaker-independent continuous speech recognition systems. It is our conviction that proper utilization of speech-specific knowledge is essential for advanced speech recognition systems. With this in mind, we have continued to make progress on the acquisition of acoustic-phonetic and lexical knowledge. We have completed the development of a continuous digit recognition system. The system was constructed to investigate the utilization of acoustic phonetic knowledge in a speech recognition system. Some of the significant development of this study includes a soft-failure procedure for lexical access, and the discovery of a set of acoustic-phonetic features for verification. We have completed a study of the constraints provided by lexical stress on word recognition. We found that lexical stress information alone can, on the average, reduce the number of word candidates from a large dictionary by more than 80%. In conjunction with this study, we successfully developed a system that automatically determines the stress pattern of a word from the acoustic signal.

  8. Acoustic Evidence for Phonologically Mismatched Speech Errors

    ERIC Educational Resources Information Center

    Gormley, Andrea

    2015-01-01

    Speech errors are generally said to accommodate to their new phonological context. This accommodation has been validated by several transcription studies. The transcription methodology is not the best choice for detecting errors at this level, however, as this type of error can be difficult to perceive. This paper presents an acoustic analysis of…

  9. Speech recognition: Acoustic, phonetic and lexical knowledge

    NASA Astrophysics Data System (ADS)

    Zue, V. W.

    1985-08-01

    During this reporting period we continued to make progress on the acquisition of acoustic-phonetic and lexical knowledge. We completed development of a continuous digit recognition system. The system was constructed to investigate the use of acoustic-phonetic knowledge in a speech recognition system. The significant achievements of this study include the development of a soft-failure procedure for lexical access and the discovery of a set of acoustic-phonetic features for verification. We completed a study of the constraints that lexical stress imposes on word recognition. We found that lexical stress information alone can, on the average, reduce the number of word candidates from a large dictionary by more than 80 percent. In conjunction with this study, we successfully developed a system that automatically determines the stress pattern of a word from the acoustic signal. We performed an acoustic study on the characteristics of nasal consonants and nasalized vowels. We have also developed recognition algorithms for nasal murmurs and nasalized vowels in continuous speech. We finished the preliminary development of a system that aligns a speech waveform with the corresponding phonetic transcription.

  10. Perceptual centres in speech - an acoustic analysis

    NASA Astrophysics Data System (ADS)

    Scott, Sophie Kerttu

    Perceptual centres, or P-centres, represent the perceptual moments of occurrence of acoustic signals - the 'beat' of a sound. P-centres underlie the perception and production of rhythm in perceptually regular speech sequences. P-centres have been modelled both in speech and non speech (music) domains. The three aims of this thesis were toatest out current P-centre models to determine which best accounted for the experimental data bto identify a candidate parameter to map P-centres onto (a local approach) as opposed to the previous global models which rely upon the whole signal to determine the P-centre the final aim was to develop a model of P-centre location which could be applied to speech and non speech signals. The first aim was investigated by a series of experiments in which a) speech from different speakers was investigated to determine whether different models could account for variation between speakers b) whether rendering the amplitude time plot of a speech signal affects the P-centre of the signal c) whether increasing the amplitude at the offset of a speech signal alters P-centres in the production and perception of speech. The second aim was carried out by a) manipulating the rise time of different speech signals to determine whether the P-centre was affected, and whether the type of speech sound ramped affected the P-centre shift b) manipulating the rise time and decay time of a synthetic vowel to determine whether the onset alteration was had more affect on P-centre than the offset manipulation c) and whether the duration of a vowel affected the P-centre, if other attributes (amplitude, spectral contents) were held constant. The third aim - modelling P-centres - was based on these results. The Frequency dependent Amplitude Increase Model of P-centre location (FAIM) was developed using a modelling protocol, the APU GammaTone Filterbank and the speech from different speakers. The P-centres of the stimuli corpus were highly predicted by attributes of

  11. Acoustic and Perceptual Characteristics of Vowels Produced during Simultaneous Communication

    ERIC Educational Resources Information Center

    Schiavetti, Nicholas; Metz, Dale Evan; Whitehead, Robert L.; Brown, Shannon; Borges, Janie; Rivera, Sara; Schultz, Christine

    2004-01-01

    This study investigated the acoustical and perceptual characteristics of vowels in speech produced during simultaneous communication (SC). Twelve normal hearing, experienced sign language users were recorded under SC and speech alone (SA) conditions speaking a set of sentences containing monosyllabic words designed for measurement of vowel…

  12. Ultrasonic speech translator and communications system

    DOEpatents

    Akerman, M.A.; Ayers, C.W.; Haynes, H.D.

    1996-07-23

    A wireless communication system undetectable by radio frequency methods for converting audio signals, including human voice, to electronic signals in the ultrasonic frequency range, transmitting the ultrasonic signal by way of acoustical pressure waves across a carrier medium, including gases, liquids, or solids, and reconverting the ultrasonic acoustical pressure waves back to the original audio signal. The ultrasonic speech translator and communication system includes an ultrasonic transmitting device and an ultrasonic receiving device. The ultrasonic transmitting device accepts as input an audio signal such as human voice input from a microphone or tape deck. The ultrasonic transmitting device frequency modulates an ultrasonic carrier signal with the audio signal producing a frequency modulated ultrasonic carrier signal, which is transmitted via acoustical pressure waves across a carrier medium such as gases, liquids or solids. The ultrasonic receiving device converts the frequency modulated ultrasonic acoustical pressure waves to a frequency modulated electronic signal, demodulates the audio signal from the ultrasonic carrier signal, and conditions the demodulated audio signal to reproduce the original audio signal at its output. 7 figs.

  13. Ultrasonic speech translator and communications system

    DOEpatents

    Akerman, M. Alfred; Ayers, Curtis W.; Haynes, Howard D.

    1996-01-01

    A wireless communication system undetectable by radio frequency methods for converting audio signals, including human voice, to electronic signals in the ultrasonic frequency range, transmitting the ultrasonic signal by way of acoustical pressure waves across a carrier medium, including gases, liquids, or solids, and reconverting the ultrasonic acoustical pressure waves back to the original audio signal. The ultrasonic speech translator and communication system (20) includes an ultrasonic transmitting device (100) and an ultrasonic receiving device (200). The ultrasonic transmitting device (100) accepts as input (115) an audio signal such as human voice input from a microphone (114) or tape deck. The ultrasonic transmitting device (100) frequency modulates an ultrasonic carrier signal with the audio signal producing a frequency modulated ultrasonic carrier signal, which is transmitted via acoustical pressure waves across a carrier medium such as gases, liquids or solids. The ultrasonic receiving device (200) converts the frequency modulated ultrasonic acoustical pressure waves to a frequency modulated electronic signal, demodulates the audio signal from the ultrasonic carrier signal, and conditions the demodulated audio signal to reproduce the original audio signal at its output (250).

  14. Acoustic evidence for phonologically mismatched speech errors.

    PubMed

    Gormley, Andrea

    2015-04-01

    Speech errors are generally said to accommodate to their new phonological context. This accommodation has been validated by several transcription studies. The transcription methodology is not the best choice for detecting errors at this level, however, as this type of error can be difficult to perceive. This paper presents an acoustic analysis of speech errors that uncovers non-accommodated or mismatch errors. A mismatch error is a sub-phonemic error that results in an incorrect surface phonology. This type of error could arise during the processing of phonological rules or they could be made at the motor level of implementation. The results of this work have important implications for both experimental and theoretical research. For experimentalists, it validates the tools used for error induction and the acoustic determination of errors free of the perceptual bias. For theorists, this methodology can be used to test the nature of the processes proposed in language production.

  15. Covert underwater acoustic communications.

    PubMed

    Ling, Jun; He, Hao; Li, Jian; Roberts, William; Stoica, Petre

    2010-11-01

    Low probability of detection (LPD) communications are conducted at a low received signal-to-noise ratio (SNR) to deter eavesdroppers to sense the presence of the transmitted signal. Successful detection at intended receiver heavily relies on the processing gain achieved by employing the direct-sequence spread-spectrum (DSSS) technique. For scenarios that lack a sufficiently low SNR to maintain LPD, another metric, referred to as low probability of interception (LPI), is of interest to protect the privacy of the transmitted information. If covert communications take place in underwater acoustic (UWA) environments, then additional challenges are present. The time-varying nature of the UWA channel prevents the employment of a long spreading waveform. Furthermore, UWA environments are frequency-selective channels with long memory, which imposes challenges to the design of the spreading waveform. In this paper, a covert UWA communication system that adopts the DSSS technique and a coherent RAKE receiver is investigated. Emphasis is placed on the design of a spreading waveform that not only accounts for the transceiver structure and frequency-selective nature of the UWA channel, but also possesses a superior LPI. The proposed techniques are evaluated using both simulated and SPACE'08 in-water experimental data.

  16. Acoustic analysis of speech under stress.

    PubMed

    Sondhi, Savita; Khan, Munna; Vijay, Ritu; Salhan, Ashok K; Chouhan, Satish

    2015-01-01

    When a person is emotionally charged, stress could be discerned in his voice. This paper presents a simplified and a non-invasive approach to detect psycho-physiological stress by monitoring the acoustic modifications during a stressful conversation. Voice database consists of audio clips from eight different popular FM broadcasts wherein the host of the show vexes the subjects who are otherwise unaware of the charade. The audio clips are obtained from real-life stressful conversations (no simulated emotions). Analysis is done using PRAAT software to evaluate mean fundamental frequency (F0) and formant frequencies (F1, F2, F3, F4) both in neutral and stressed state. Results suggest that F0 increases with stress; however, formant frequency decreases with stress. Comparison of Fourier and chirp spectra of short vowel segment shows that for relaxed speech, the two spectra are similar; however, for stressed speech, they differ in the high frequency range due to increased pitch modulation. PMID:26558301

  17. System And Method For Characterizing Voiced Excitations Of Speech And Acoustic Signals, Removing Acoustic Noise From Speech, And Synthesizi

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2006-04-25

    The present invention is a system and method for characterizing human (or animate) speech voiced excitation functions and acoustic signals, for removing unwanted acoustic noise which often occurs when a speaker uses a microphone in common environments, and for synthesizing personalized or modified human (or other animate) speech upon command from a controller. A low power EM sensor is used to detect the motions of windpipe tissues in the glottal region of the human speech system before, during, and after voiced speech is produced by a user. From these tissue motion measurements, a voiced excitation function can be derived. Further, the excitation function provides speech production information to enhance noise removal from human speech and it enables accurate transfer functions of speech to be obtained. Previously stored excitation and transfer functions can be used for synthesizing personalized or modified human speech. Configurations of EM sensor and acoustic microphone systems are described to enhance noise cancellation and to enable multiple articulator measurements.

  18. Intelligibility of Speech Produced during Simultaneous Communication

    ERIC Educational Resources Information Center

    Whitehead, Robert L.; Schiavetti, Nicholas; MacKenzie, Douglas J.; Metz, Dale Evan

    2004-01-01

    This study investigated the overall intelligibility of speech produced during simultaneous communication (SC). Four hearing, experienced sign language users were recorded under SC and speech alone (SA) conditions speaking Boothroyd's (1985) forced-choice phonetic contrast material designed for measurement of speech intelligibility. Twelve…

  19. Communicating by Language: The Speech Process.

    ERIC Educational Resources Information Center

    House, Arthur S., Ed.

    This document reports on a conference focused on speech problems. The main objective of these discussions was to facilitate a deeper understanding of human communication through interaction of conference participants with colleagues in other disciplines. Topics discussed included speech production, feedback, speech perception, and development of…

  20. Acoustic Study of Acted Emotions in Speech

    NASA Astrophysics Data System (ADS)

    Wang, Rong

    An extensive set of carefully recorded utterances provided a speech database for investigating acoustic correlates among eight emotional states. Four professional actors and four professional actresses simulated the emotional states of joy, conversation, nervousness, anger, sadness, hate, fear, and depression. The values of 14 acoustic parameters were extracted from analyses of the simulated portrayals. Normalization of the parameters was made to reduce the talker-dependence. Correlates of emotion were investigated by means of principal components analysis. Sadness and depression were found to be "ambiguous" with respect to each other, but "unique" with respect to joy and anger in the principal components space. Joy, conversation, nervousness, anger, hate, and fear did not separate well in the space and so exhibited ambiguity with respect to one another. The different talkers expressed joy, anger, sadness, and depression more consistently than the other four emotions. The analysis results were compared with the results of a subjective study using the same speech database and considerable consistency between the two was found.

  1. Acoustic differences among casual, conversational, and read speech

    NASA Astrophysics Data System (ADS)

    Pinnow, DeAnna

    Speech is a complex behavior that allows speakers to use many variations to satisfy the demands connected with multiple speaking environments. Speech research typically obtains speech samples in a controlled laboratory setting using read material, yet anecdotal observations of such speech, particularly from talkers with a speech and language impairment, have identified a "performance" effect in the produced speech which masks the characteristics of impaired speech outside of the lab (Goberman, Recker, & Parveen, 2010). The aim of the current study was to investigate acoustic differences among laboratory read, laboratory conversational, and casual speech through well-defined speech tasks in the laboratory and in talkers' natural environments. Eleven healthy research participants performed lab recording tasks (19 read sentences and a dialogue about their life) and collected natural-environment recordings of themselves over 3-day periods using portable recorders. Segments were analyzed for articulatory, voice, and prosodic acoustic characteristics using computer software and hand counting. The current study results indicate that lab-read speech was significantly different from casual speech: greater articulation range, improved voice quality measures, lower speech rate, and lower mean pitch. One implication of the results is that different laboratory techniques may be beneficial in obtaining speech samples that are more like casual speech, thus making it easier to correctly analyze abnormal speech characteristics with fewer errors.

  2. Education in the 80's: Speech Communication.

    ERIC Educational Resources Information Center

    Friedrich, Gustav W., Ed.

    Taken together, the 20 chapters in this book provide many suggestions, predictions, alternatives, innovations, and improvements in the speech communication curriculum that can be either undertaken or accomplished during the 1980s. The first five chapters speculate positively about the future of speech communication instruction in five of its most…

  3. [Thematic Issue: Career Trends in Speech Communication.

    ERIC Educational Resources Information Center

    Hall, Robert, Ed.

    To analyze historical trends in job opportunities available to speech communications graduates, a content analysis was conducted of Speech Communication Association bulletins over a ten-year period (1967 to 1977). All bulletin listings were analyzed for state, type of institution, rank, areas of specialization, job requirements, and salary range.…

  4. Acoustics in Halls for Speech and Music

    NASA Astrophysics Data System (ADS)

    Gade, Anders C.

    This chapter deals specifically with concepts, tools, and architectural variables of importance when designing auditoria for speech and music. The focus will be on cultivating the useful components of the sound in the room rather than on avoiding noise from outside or from installations, which is dealt with in Chap. 11. The chapter starts by presenting the subjective aspects of the room acoustic experience according to consensus at the time of writing. Then follows a description of their objective counterparts, the objective room acoustic parameters, among which the classical reverberation time measure is only one of many, but still of fundamental value. After explanations on how these parameters can be measured and predicted during the design phase, the remainder of the chapter deals with how the acoustic properties can be controlled by the architectural design of auditoria. This is done by presenting the influence of individual design elements as well as brief descriptions of halls designed for specific purposes, such as drama, opera, and symphonic concerts. Finally, some important aspects of loudspeaker installations in auditoria are briefly touched upon.

  5. The acoustic properties of bilingual infant-directed speech.

    PubMed

    Danielson, D Kyle; Seidl, Amanda; Onishi, Kristine H; Alamian, Golnoush; Cristia, Alejandrina

    2014-02-01

    Does the acoustic input for bilingual infants equal the conjunction of the input heard by monolinguals of each separate language? The present letter tackles this question, focusing on maternal speech addressed to 11-month-old infants, on the cusp of perceptual attunement. The acoustic characteristics of the point vowels /a,i,u/ were measured in the spontaneous infant-directed speech of French-English bilingual mothers, as well as in the speech of French and English monolingual mothers. Bilingual caregivers produced their two languages with acoustic prosodic separation equal to that of the monolinguals, while also conveying distinct spectral characteristics of the point vowels in their two languages.

  6. Alternative Speech Communication System for Persons with Severe Speech Disorders

    NASA Astrophysics Data System (ADS)

    Selouani, Sid-Ahmed; Sidi Yakoub, Mohammed; O'Shaughnessy, Douglas

    2009-12-01

    Assistive speech-enabled systems are proposed to help both French and English speaking persons with various speech disorders. The proposed assistive systems use automatic speech recognition (ASR) and speech synthesis in order to enhance the quality of communication. These systems aim at improving the intelligibility of pathologic speech making it as natural as possible and close to the original voice of the speaker. The resynthesized utterances use new basic units, a new concatenating algorithm and a grafting technique to correct the poorly pronounced phonemes. The ASR responses are uttered by the new speech synthesis system in order to convey an intelligible message to listeners. Experiments involving four American speakers with severe dysarthria and two Acadian French speakers with sound substitution disorders (SSDs) are carried out to demonstrate the efficiency of the proposed methods. An improvement of the Perceptual Evaluation of the Speech Quality (PESQ) value of 5% and more than 20% is achieved by the speech synthesis systems that deal with SSD and dysarthria, respectively.

  7. Clear Speech Variants: An Acoustic Study in Parkinson's Disease

    ERIC Educational Resources Information Center

    Lam, Jennifer; Tjaden, Kris

    2016-01-01

    Purpose: The authors investigated how different variants of clear speech affect segmental and suprasegmental acoustic measures of speech in speakers with Parkinson's disease and a healthy control group. Method: A total of 14 participants with Parkinson's disease and 14 control participants served as speakers. Each speaker produced 18 different…

  8. Acoustic Markers of Prosodic Boundaries in Spanish Spontaneous Alaryngeal Speech

    ERIC Educational Resources Information Center

    Cuenca, M. H.; Barrio, M. M.

    2010-01-01

    Prosodic information aids segmentation of the continuous speech signal and thereby facilitates auditory speech processing. Durational and pitch variations are prosodic cues especially necessary to convey prosodic boundaries, but alaryngeal speakers have inconsistent control over acoustic parameters such as F0 and duration, being as a result noisy…

  9. SPEECH--MAN'S NATURAL COMMUNICATION.

    ERIC Educational Resources Information Center

    DUDLEY, HOMER; AND OTHERS

    SESSION 63 OF THE 1967 INSTITUTE OF ELECTRICAL AND ELECTRONIC ENGINEERS INTERNATIONAL CONVENTION BROUGHT TOGETHER SEVEN DISTINGUISHED MEN WORKING IN FIELDS RELEVANT TO LANGUAGE. THEIR TOPICS INCLUDED ORIGIN AND EVOLUTION OF SPEECH AND LANGUAGE, LANGUAGE AND CULTURE, MAN'S PHYSIOLOGICAL MECHANISMS FOR SPEECH, LINGUISTICS, AND TECHNOLOGY AND…

  10. Communication acoustics in Bell Labs

    NASA Astrophysics Data System (ADS)

    Flanagan, J. L.

    2001-05-01

    Communication aoustics has been a central theme in Bell Labs research since its inception. Telecommunication serves human information exchange. And, humans favor spoken language as a principal mode. The atmospheric medium typically provides the link between articulation and hearing. Creation, control and detection of sound, and the human's facility for generation and perception are basic ingredients of telecommunication. Electronics technology of the 1920s ushered in great advances in communication at a distance, a strong economical impetus being to overcome bandwidth limitations of wireline and cable. Early research established criteria for speech transmission with high quality and intelligibility. These insights supported exploration of means for efficient transmission-obtaining the greatest amount of speech information over a given bandwidth. Transoceanic communication was initiated by undersea cables for telegraphy. But these long cables exhibited very limited bandwidth (order of few hundred Hz). The challenge of sending voice across the oceans spawned perhaps the best known speech compression technique of history-the Vocoder, which parametrized the signal for transmission in about 300 Hz bandwidth, one-tenth that required for the typical waveform channel. Quality and intelligibility were grave issues (and they still are). At the same time parametric representation offered possibilities for encryption and privacy inside a traditional voice bandwidth. Confidential conversations between Roosevelt and Churchill during World War II were carried over high-frequency radio by an encrypted vocoder system known as Sigsaly. Major engineering advances in the late 1940s and early 1950s moved telecommunications into a new regime-digital technology. These key advances were at least three: (i) new understanding of time-discrete (sampled) representation of signals, (ii) digital computation (especially binary based), and (iii) evolving capabilities in microelectronics that

  11. Study of acoustic correlates associate with emotional speech

    NASA Astrophysics Data System (ADS)

    Yildirim, Serdar; Lee, Sungbok; Lee, Chul Min; Bulut, Murtaza; Busso, Carlos; Kazemzadeh, Ebrahim; Narayanan, Shrikanth

    2004-10-01

    This study investigates the acoustic characteristics of four different emotions expressed in speech. The aim is to obtain detailed acoustic knowledge on how a speech signal is modulated by changes from neutral to a certain emotional state. Such knowledge is necessary for automatic emotion recognition and classification and emotional speech synthesis. Speech data obtained from two semi-professional actresses are analyzed and compared. Each subject produces 211 sentences with four different emotions; neutral, sad, angry, happy. We analyze changes in temporal and acoustic parameters such as magnitude and variability of segmental duration, fundamental frequency and the first three formant frequencies as a function of emotion. Acoustic differences among the emotions are also explored with mutual information computation, multidimensional scaling and acoustic likelihood comparison with normal speech. Results indicate that speech associated with anger and happiness is characterized by longer duration, shorter interword silence, higher pitch and rms energy with wider ranges. Sadness is distinguished from other emotions by lower rms energy and longer interword silence. Interestingly, the difference in formant pattern between [happiness/anger] and [neutral/sadness] are better reflected in back vowels such as /a/(/father/) than in front vowels. Detailed results on intra- and interspeaker variability will be reported.

  12. Acoustic assessment of speech privacy curtains in two nursing units

    PubMed Central

    Pope, Diana S.; Miller-Klein, Erik T.

    2016-01-01

    Hospitals have complex soundscapes that create challenges to patient care. Extraneous noise and high reverberation rates impair speech intelligibility, which leads to raised voices. In an unintended spiral, the increasing noise may result in diminished speech privacy, as people speak loudly to be heard over the din. The products available to improve hospital soundscapes include construction materials that absorb sound (acoustic ceiling tiles, carpet, wall insulation) and reduce reverberation rates. Enhanced privacy curtains are now available and offer potential for a relatively simple way to improve speech privacy and speech intelligibility by absorbing sound at the hospital patient's bedside. Acoustic assessments were performed over 2 days on two nursing units with a similar design in the same hospital. One unit was built with the 1970s’ standard hospital construction and the other was newly refurbished (2013) with sound-absorbing features. In addition, we determined the effect of an enhanced privacy curtain versus standard privacy curtains using acoustic measures of speech privacy and speech intelligibility indexes. Privacy curtains provided auditory protection for the patients. In general, that protection was increased by the use of enhanced privacy curtains. On an average, the enhanced curtain improved sound absorption from 20% to 30%; however, there was considerable variability, depending on the configuration of the rooms tested. Enhanced privacy curtains provide measureable improvement to the acoustics of patient rooms but cannot overcome larger acoustic design issues. To shorten reverberation time, additional absorption, and compact and more fragmented nursing unit floor plate shapes should be considered. PMID:26780959

  13. Acoustic assessment of speech privacy curtains in two nursing units.

    PubMed

    Pope, Diana S; Miller-Klein, Erik T

    2016-01-01

    Hospitals have complex soundscapes that create challenges to patient care. Extraneous noise and high reverberation rates impair speech intelligibility, which leads to raised voices. In an unintended spiral, the increasing noise may result in diminished speech privacy, as people speak loudly to be heard over the din. The products available to improve hospital soundscapes include construction materials that absorb sound (acoustic ceiling tiles, carpet, wall insulation) and reduce reverberation rates. Enhanced privacy curtains are now available and offer potential for a relatively simple way to improve speech privacy and speech intelligibility by absorbing sound at the hospital patient's bedside. Acoustic assessments were performed over 2 days on two nursing units with a similar design in the same hospital. One unit was built with the 1970s' standard hospital construction and the other was newly refurbished (2013) with sound-absorbing features. In addition, we determined the effect of an enhanced privacy curtain versus standard privacy curtains using acoustic measures of speech privacy and speech intelligibility indexes. Privacy curtains provided auditory protection for the patients. In general, that protection was increased by the use of enhanced privacy curtains. On an average, the enhanced curtain improved sound absorption from 20% to 30%; however, there was considerable variability, depending on the configuration of the rooms tested. Enhanced privacy curtains provide measureable improvement to the acoustics of patient rooms but cannot overcome larger acoustic design issues. To shorten reverberation time, additional absorption, and compact and more fragmented nursing unit floor plate shapes should be considered. PMID:26780959

  14. Acoustic assessment of speech privacy curtains in two nursing units.

    PubMed

    Pope, Diana S; Miller-Klein, Erik T

    2016-01-01

    Hospitals have complex soundscapes that create challenges to patient care. Extraneous noise and high reverberation rates impair speech intelligibility, which leads to raised voices. In an unintended spiral, the increasing noise may result in diminished speech privacy, as people speak loudly to be heard over the din. The products available to improve hospital soundscapes include construction materials that absorb sound (acoustic ceiling tiles, carpet, wall insulation) and reduce reverberation rates. Enhanced privacy curtains are now available and offer potential for a relatively simple way to improve speech privacy and speech intelligibility by absorbing sound at the hospital patient's bedside. Acoustic assessments were performed over 2 days on two nursing units with a similar design in the same hospital. One unit was built with the 1970s' standard hospital construction and the other was newly refurbished (2013) with sound-absorbing features. In addition, we determined the effect of an enhanced privacy curtain versus standard privacy curtains using acoustic measures of speech privacy and speech intelligibility indexes. Privacy curtains provided auditory protection for the patients. In general, that protection was increased by the use of enhanced privacy curtains. On an average, the enhanced curtain improved sound absorption from 20% to 30%; however, there was considerable variability, depending on the configuration of the rooms tested. Enhanced privacy curtains provide measureable improvement to the acoustics of patient rooms but cannot overcome larger acoustic design issues. To shorten reverberation time, additional absorption, and compact and more fragmented nursing unit floor plate shapes should be considered.

  15. Speech Communication as a Required University Course.

    ERIC Educational Resources Information Center

    Friedrich, Gustav W.

    1982-01-01

    Argues that a speech communication course should be a required course because speaking and listening are important basic skills which require refinement during a student's college/university years under the guidance of communication professionals. Advises departments to consider the resources (people and material) necessary to develop and…

  16. Acoustic communication in plant-animal interactions.

    PubMed

    Schöner, Michael G; Simon, Ralph; Schöner, Caroline R

    2016-08-01

    Acoustic communication is widespread and well-studied in animals but has been neglected in other organisms such as plants. However, there is growing evidence for acoustic communication in plant-animal interactions. While knowledge about active acoustic signalling in plants (i.e. active sound production) is still in its infancy, research on passive acoustic signalling (i.e. reflection of animal sounds) revealed that bat-dependent plants have adapted to the bats' echolocation systems by providing acoustic reflectors to attract their animal partners. Understanding the proximate mechanisms and ultimate causes of acoustic communication will shed light on an underestimated dimension of information transfer between plants and animals. PMID:27423052

  17. Acoustic communication in plant-animal interactions.

    PubMed

    Schöner, Michael G; Simon, Ralph; Schöner, Caroline R

    2016-08-01

    Acoustic communication is widespread and well-studied in animals but has been neglected in other organisms such as plants. However, there is growing evidence for acoustic communication in plant-animal interactions. While knowledge about active acoustic signalling in plants (i.e. active sound production) is still in its infancy, research on passive acoustic signalling (i.e. reflection of animal sounds) revealed that bat-dependent plants have adapted to the bats' echolocation systems by providing acoustic reflectors to attract their animal partners. Understanding the proximate mechanisms and ultimate causes of acoustic communication will shed light on an underestimated dimension of information transfer between plants and animals.

  18. Evaluating a topographical mapping from speech acoustics to tongue positions

    SciTech Connect

    Hogden, J.; Heard, M.

    1995-05-01

    The {ital continuity} {ital mapping} algorithm---a procedure for learning to recover the relative positions of the articulators from speech signals---is evaluated using human speech data. The advantage of continuity mapping is that it is an unsupervised algorithm; that is, it can potentially be trained to make a mapping from speech acoustics to speech articulation without articulator measurements. The procedure starts by vector quantizing short windows of a speech signal so that each window is represented (encoded) by a single number. Next, multidimensional scaling is used to map quantization codes that were temporally close in the encoded speech to nearby points in a {ital continuity} {ital map}. Since speech sounds produced sufficiently close together in time must have been produced by similar articulator configurations, and speech sounds produced close together in time are close to each other in the continuity map, sounds produced by similar articulator positions should be mapped to similar positions in the continuity map. The data set used for evaluating the continuity mapping algorithm is comprised of simultaneously collected articulator and acoustic measurements made using an electromagnetic midsagittal articulometer on a human subject. Comparisons between measured articulator positions and those recovered using continuity mapping will be presented.

  19. Methods and apparatus for non-acoustic speech characterization and recognition

    SciTech Connect

    Holzrichter, J.F.

    1999-12-21

    By simultaneously recording EM wave reflections and acoustic speech information, the positions and velocities of the speech organs as speech is articulated can be defined for each acoustic speech unit. Well defined time frames and feature vectors describing the speech, to the degree required, can be formed. Such feature vectors can uniquely characterize the speech unit being articulated each time frame. The onset of speech, rejection of external noise, vocalized pitch periods, articulator conditions, accurate timing, the identification of the speaker, acoustic speech unit recognition, and organ mechanical parameters can be determined.

  20. Methods and apparatus for non-acoustic speech characterization and recognition

    DOEpatents

    Holzrichter, John F.

    1999-01-01

    By simultaneously recording EM wave reflections and acoustic speech information, the positions and velocities of the speech organs as speech is articulated can be defined for each acoustic speech unit. Well defined time frames and feature vectors describing the speech, to the degree required, can be formed. Such feature vectors can uniquely characterize the speech unit being articulated each time frame. The onset of speech, rejection of external noise, vocalized pitch periods, articulator conditions, accurate timing, the identification of the speaker, acoustic speech unit recognition, and organ mechanical parameters can be determined.

  1. Investigation of the optimum acoustical conditions for speech using auralization

    NASA Astrophysics Data System (ADS)

    Yang, Wonyoung; Hodgson, Murray

    2001-05-01

    Speech intelligibility is mainly affected by reverberation and by signal-to-noise level difference, the difference between the speech-signal and background-noise levels at a receiver. An important question for the design of rooms for speech (e.g., classrooms) is, what are the optimal values of these factors? This question has been studied experimentally and theoretically. Experimental studies found zero optimal reverberation time, but theoretical predictions found nonzero reverberation times. These contradictory results are partly caused by the different ways of accounting for background noise. Background noise sources and their locations inside the room are the most detrimental factors in speech intelligibility. However, noise levels also interact with reverberation in rooms. In this project, two major room-acoustical factors for speech intelligibility were controlled using speech and noise sources of known relative output levels located in a virtual room with known reverberation. Speech intelligibility test signals were played in the virtual room and auralized for listeners. The Modified Rhyme Test (MRT) and babble noise were used to measure subjective speech intelligibility quality. Optimal reverberation times, and the optimal values of other speech intelligibility metrics, for normal-hearing people and for hard-of-hearing people, were identified and compared.

  2. Acoustic Speech Analysis Of Wayang Golek Puppeteer

    NASA Astrophysics Data System (ADS)

    Hakim, Faisal Abdul; Mandasari, Miranti Indar; Sarwono, Joko

    2010-12-01

    Active disguising speech is one problem to be taken into account in forensic speaker verification or identification processes. The verification processes are usually carried out by comparison between unknown samples and known samples. Active disguising can be occurred on both samples. To simulate the condition of speech disguising, voices of Wayang Golek Puppeteer were used. It is assumed that wayang golek puppeteer is a master of disguise. He can manipulate his voice into many different types of character's voices. This paper discusses the speech characteristics of 2 puppeteers. Comparison was made between the voices of puppeteer's habitual voice with his manipulated voice.

  3. The Carolinas Speech Communication Annual, 1996.

    ERIC Educational Resources Information Center

    McKinney, Bruce C., Ed.

    1996-01-01

    This 1996 issue of the "Carolinas Speech Communication Annual" contains the following articles: "Rhetoric in the Second Sophistic, Medieval, and Renaissance Periods: Implications for Pedagogy" (Omar J. Swartz and Chris Bachelder); "Thou Art Damned: Cursing as a Rhetorical Strategy of the Accusers in the Salem Witch Trials" (Colleen E. Kelley);…

  4. Implications of Empathy Research for Speech Communication.

    ERIC Educational Resources Information Center

    Jackson, Dale Myers

    In this inductive study, a body of empathy-related research is reviewed for the purpose of identifying theoretical and pedagogical implications for a college-level, introductory course in speech communication. The results of studies conducted primarily in the fields of psychotherapy and counseling are summarized. Evidence which places empathy at…

  5. Conceptual Divisions Within the Speech Communication Association.

    ERIC Educational Resources Information Center

    Stone, John D.

    This paper specifically discusses the differing methods by which certain conceptual divisions in the field of speech communication have been determined. Classifying the various conceptual approaches as (1) definitional analyses, (2) structural approaches, (3) research classifications, and (4) approaches reflecting a combination of these three, the…

  6. On the State of Speech Communication Research.

    ERIC Educational Resources Information Center

    Sillars, Malcolm O.

    The discipline of speech communication was established early in the twentieth century. After its subject matter was defined and separated from the field of English by Herbert Wichelns's "The Literary Criticism of Oratory," departments were established and a separate organization--The National Association of Teachers of Public Speaking--was formed.…

  7. Do 6-Month-Olds Understand That Speech Can Communicate?

    ERIC Educational Resources Information Center

    Vouloumanos, Athena; Martin, Alia; Onishi, Kristine H.

    2014-01-01

    Adults and 12-month-old infants recognize that even unfamiliar speech can communicate information between third parties, suggesting that they can separate the communicative function of speech from its lexical content. But do infants recognize that speech can communicate due to their experience understanding and producing language, or do they…

  8. Music, Speech and Hearing: A Course in Descriptive Acoustics.

    ERIC Educational Resources Information Center

    Strong, William J.; Dudley, J. Duane

    A general undergraduate education course in descriptive acoustics for students in music, speech, and environmental disciplines is described. Student background, course philosophy, and course format are presented. Resource materials, which include books, a study guide, films, audio tapes, demonstrations, walk-in laboratories, and an examination…

  9. System and method for characterizing voiced excitations of speech and acoustic signals, removing acoustic noise from speech, and synthesizing speech

    DOEpatents

    Burnett, Greg C.; Holzrichter, John F.; Ng, Lawrence C.

    2002-01-01

    Low power EM waves are used to detect motions of vocal tract tissues of the human speech system before, during, and after voiced speech. A voiced excitation function is derived. The excitation function provides speech production information to enhance speech characterization and to enable noise removal from human speech.

  10. Spatial acoustic signal processing for immersive communication

    NASA Astrophysics Data System (ADS)

    Atkins, Joshua

    Computing is rapidly becoming ubiquitous as users expect devices that can augment and interact naturally with the world around them. In these systems it is necessary to have an acoustic front-end that is able to capture and reproduce natural human communication. Whether the end point is a speech recognizer or another human listener, the reduction of noise, reverberation, and acoustic echoes are all necessary and complex challenges. The focus of this dissertation is to provide a general method for approaching these problems using spherical microphone and loudspeaker arrays.. In this work, a theory of capturing and reproducing three-dimensional acoustic fields is introduced from a signal processing perspective. In particular, the decomposition of the spatial part of the acoustic field into an orthogonal basis of spherical harmonics provides not only a general framework for analysis, but also many processing advantages. The spatial sampling error limits the upper frequency range with which a sound field can be accurately captured or reproduced. In broadband arrays, the cost and complexity of using multiple transducers is an issue. This work provides a flexible optimization method for determining the location of array elements to minimize the spatial aliasing error. The low frequency array processing ability is also limited by the SNR, mismatch, and placement error of transducers. To address this, a robust processing method is introduced and used to design a reproduction system for rendering over arbitrary loudspeaker arrays or binaurally over headphones. In addition to the beamforming problem, the multichannel acoustic echo cancellation (MCAEC) issue is also addressed. A MCAEC must adaptively estimate and track the constantly changing loudspeaker-room-microphone response to remove the sound field presented over the loudspeakers from that captured by the microphones. In the multichannel case, the system is overdetermined and many adaptive schemes fail to converge to

  11. Amplitude Modulations of Acoustic Communication Signals

    NASA Astrophysics Data System (ADS)

    Turesson, Hjalmar K.

    2011-12-01

    In human speech, amplitude modulations at 3 -- 8 Hz are important for discrimination and detection. Two different neurophysiological theories have been proposed to explain this effect. The first theory proposes that, as a consequence of neocortical synaptic dynamics, signals that are amplitude modulated at 3 -- 8 Hz are propagated better than un-modulated signals, or signals modulated above 8 Hz. This suggests that neural activity elicited by vocalizations modulated at 3 -- 8 Hz is optimally transmitted, and the vocalizations better discriminated and detected. The second theory proposes that 3 -- 8 Hz amplitude modulations interact with spontaneous neocortical oscillations. Specifically, vocalizations modulated at 3 -- 8 Hz entrain local populations of neurons, which in turn, modulate the amplitude of high frequency gamma oscillations. This suggests that vocalizations modulated at 3 -- 8 Hz should induce stronger cross-frequency coupling. Similar to human speech, we found that macaque monkey vocalizations also are amplitude modulated between 3 and 8 Hz. Humans and macaque monkeys share similarities in vocal production, implying that the auditory systems subserving perception of acoustic communication signals also share similarities. Based on the similarities between human speech and macaque monkey vocalizations, we addressed how amplitude modulated vocalizations are processed in the auditory cortex of macaque monkeys, and what behavioral relevance modulations may have. Recording single neuron activity, as well as, the activity of local populations of neurons allowed us to test both of the neurophysiological theories presented above. We found that single neuron responses to vocalizations amplitude modulated at 3 -- 8 Hz resulted in better stimulus discrimination than vocalizations lacking 3 -- 8 Hz modulations, and that the effect most likely was mediated by synaptic dynamics. In contrast, we failed to find support for the oscillation-based model proposing a

  12. Speech recognition: Acoustic-phonetic knowledge acquisition and representation

    NASA Astrophysics Data System (ADS)

    Zue, Victor W.

    1988-09-01

    The long-term research goal is to develop and implement speaker-independent continuous speech recognition systems. It is believed that the proper utilization of speech-specific knowledge is essential for such advanced systems. This research is thus directed toward the acquisition, quantification, and representation, of acoustic-phonetic and lexical knowledge, and the application of this knowledge to speech recognition algorithms. In addition, we are exploring new speech recognition alternatives based on artificial intelligence and connectionist techniques. We developed a statistical model for predicting the acoustic realization of stop consonants in various positions in the syllable template. A unification-based grammatical formalism was developed for incorporating this model into the lexical access algorithm. We provided an information-theoretic justification for the hierarchical structure of the syllable template. We analyzed segmented duration for vowels and fricatives in continuous speech. Based on contextual information, we developed durational models for vowels and fricatives that account for over 70 percent of the variance, using data from multiple, unknown speakers. We rigorously evaluated the ability of human spectrogram readers to identify stop consonants spoken by many talkers and in a variety of phonetic contexts. Incorporating the declarative knowledge used by the readers, we developed a knowledge-based system for stop identification. We achieved comparable system performance to that to the readers.

  13. Acoustic Characteristics of Ataxic Speech in Japanese Patients with Spinocerebellar Degeneration (SCD)

    ERIC Educational Resources Information Center

    Ikui, Yukiko; Tsukuda, Mamoru; Kuroiwa, Yoshiyuki; Koyano, Shigeru; Hirose, Hajime; Taguchi, Takahide

    2012-01-01

    Background: In English- and German-speaking countries, ataxic speech is often described as showing scanning based on acoustic impressions. Although the term "scanning" is generally considered to represent abnormal speech features including prosodic excess or insufficiency, any precise acoustic analysis of ataxic speech has not been performed in…

  14. Segment-based acoustic models for continuous speech recognition

    NASA Astrophysics Data System (ADS)

    Ostendorf, Mari; Rohlicek, J. R.

    1994-02-01

    In work, we are interested in the problem of large vocabulary, speaker-independent continuous speech recognition, and primarily in the acoustic modeling component of this problem. In developing acoustic models for speech recognition, we have conflicting goals. On one hand, the models should be robust to inter- and intra-speaker variability, to the use of a different vocabulary in recognition than in training, and to the effects of moderately noisy environments. In order to accomplish this, we need to model gross features and global trends. On the other hand, the models must be sensitive and detailed enough to detect fine acoustic differences between similar words in a large vocabulary task. To answer these opposing demands requires improvements in acoustic modeling at several levels: the frame level (e.g. signal processing), the phoneme level (e.g. modeling feature dynamics), and the utterance level (e.g. defining a structural context for representing the intra-utterance dependence across phonemes). This project address the problem of acoustic modeling specifically focusing on modeling at the segment level and above.

  15. A method for determining internal noise criteria based on practical speech communication applied to helicopters

    NASA Technical Reports Server (NTRS)

    Sternfeld, H., Jr.; Doyle, L. B.

    1978-01-01

    The relationship between the internal noise environment of helicopters and the ability of personnel to understand commands and instructions was studied. A test program was conducted to relate speech intelligibility to a standard measurement called Articulation Index. An acoustical simulator was used to provide noise environments typical of Army helicopters. Speech material (command sentences and phonetically balanced word lists) were presented at several voice levels in each helicopter environment. Recommended helicopter internal noise criteria, based on speech communication, were derived and the effectiveness of hearing protection devices were evaluated.

  16. Predicting the intelligibility of deaf children's speech from acoustic measures

    NASA Astrophysics Data System (ADS)

    Uchanski, Rosalie M.; Geers, Ann E.; Brenner, Christine M.; Tobey, Emily A.

    2001-05-01

    A weighted combination of speech-acoustic measures may provide an objective assessment of speech intelligibility in deaf children that could be used to evaluate the benefits of sensory aids and rehabilitation programs. This investigation compared the accuracy of two different approaches, multiple linear regression and a simple neural net. These two methods were applied to identical sets of acoustic measures, including both segmental (e.g., voice-onset times of plosives, spectral moments of fricatives, second formant frequencies of vowels) and suprasegmental measures (e.g., sentence duration, number and frequency of intersentence pauses). These independent variables were obtained from digitized recordings of deaf children's imitations of 11 simple sentences. The dependent measure was the percentage of spoken words from the 36 McGarr Sentences understood by groups of naive listeners. The two predictive methods were trained on speech measures obtained from 123 out of 164 8- and 9-year-old deaf children who used cochlear implants. Then, predictions were obtained using speech measures from the remaining 41 children. Preliminary results indicate that multiple linear regression is a better predictor of intelligibility than the neural net, accounting for 79% as opposed to 65% of the variance in the data. [Work supported by NIH.

  17. Tongue-Palate Contact Pressure, Oral Air Pressure, and Acoustics of Clear Speech

    ERIC Educational Resources Information Center

    Searl, Jeff; Evitts, Paul M.

    2013-01-01

    Purpose: The authors compared articulatory contact pressure (ACP), oral air pressure (Po), and speech acoustics for conversational versus clear speech. They also assessed the relationship of these measures to listener perception. Method: Twelve adults with normal speech produced monosyllables in a phrase using conversational and clear speech.…

  18. Multichannel spatial auditory display for speech communications

    NASA Technical Reports Server (NTRS)

    Begault, D. R.; Erbe, T.; Wenzel, E. M. (Principal Investigator)

    1994-01-01

    A spatial auditory display for multiple speech communications was developed at NASA/Ames Research Center. Input is spatialized by the use of simplified head-related transfer functions, adapted for FIR filtering on Motorola 56001 digital signal processors. Hardware and firmware design implementations are overviewed for the initial prototype developed for NASA-Kennedy Space Center. An adaptive staircase method was used to determine intelligibility levels of four-letter call signs used by launch personnel at NASA against diotic speech babble. Spatial positions at 30 degrees azimuth increments were evaluated. The results from eight subjects showed a maximum intelligibility improvement of about 6-7 dB when the signal was spatialized to 60 or 90 degrees azimuth positions.

  19. How Speech Communication Training Interfaces with Public Relations Training.

    ERIC Educational Resources Information Center

    Bosley, Phyllis B.

    Speech communication training is a valuable asset for those entering the public relations (PR) field. This notion is reinforced by the 1987 "Design for Undergraduate Public Relations Education," a guide for implementing speech communication courses within a public relations curriculum, and also in the incorporation of oral communication training…

  20. Speech Intelligibility of Two Voice Output Communication Aids.

    ERIC Educational Resources Information Center

    Kannenberg, Patricia; And Others

    1988-01-01

    The intelligibility of two voice-output communication aids ("Personal Communicator" and "SpeechPAC'") was evaluated by presenting synthesized words and sentences to 20 listeners. Analysis of listener transcriptions revealed significantly higher intelligibility scores for the "Personal Communicator" compared to the "SpeechPAC" and higher scores for…

  1. Adding articulatory features to acoustic features for automatic speech recognition

    SciTech Connect

    Zlokarnik, I.

    1995-05-01

    A hidden-Markov-model (HMM) based speech recognition system was evaluated that makes use of simultaneously recorded acoustic and articulatory data. The articulatory measurements were gathered by means of electromagnetic articulography and describe the movement of small coils fixed to the speakers` tongue and jaw during the production of German V{sub 1}CV{sub 2} sequences [P. Hoole and S. Gfoerer, J. Acoust. Soc. Am. Suppl. 1 {bold 87}, S123 (1990)]. Using the coordinates of the coil positions as an articulatory representation, acoustic and articulatory features were combined to make up an acoustic--articulatory feature vector. The discriminant power of this combined representation was evaluated for two subjects on a speaker-dependent isolated word recognition task. When the articulatory measurements were used both for training and testing the HMMs, the articulatory representation was capable of reducing the error rate of comparable acoustic-based HMMs by a relative percentage of more than 60%. In a separate experiment, the articulatory movements during the testing phase were estimated using a multilayer perceptron that performed an acoustic-to-articulatory mapping. Under these more realistic conditions, when articulatory measurements are only available during the training, the error rate could be reduced by a relative percentage of 18% to 25%.

  2. Private and Inner Speech and the Regulation of Social Speech Communication

    ERIC Educational Resources Information Center

    San Martin Martinez, Conchi; Boada i Calbet, Humbert; Feigenbaum, Peter

    2011-01-01

    To further investigate the possible regulatory role of private and inner speech in the context of referential social speech communications, a set of clear and systematically applied measures is needed. This study addresses this need by introducing a rigorous method for identifying private speech and certain sharply defined instances of inaudible…

  3. Automatic speech recognition technology development at ITT Defense Communications Division

    NASA Technical Reports Server (NTRS)

    White, George M.

    1977-01-01

    An assessment of the applications of automatic speech recognition to defense communication systems is presented. Future research efforts include investigations into the following areas: (1) dynamic programming; (2) recognition of speech degraded by noise; (3) speaker independent recognition; (4) large vocabulary recognition; (5) word spotting and continuous speech recognition; and (6) isolated word recognition.

  4. Harvey Fletcher's role in the creation of communication acoustics.

    PubMed

    Allen, J B

    1996-04-01

    As the reader might appreciate after reading Fletcher's 1953 views, in 1918 Fletcher had taken on the toughest problem of all: to quantify and model how we hear and understand speech. This understanding allowed AT&T Bell Labs engineers to develop the necessary specifications of what was to become the largest telephone network in the world. The problems that Fletcher and his colleagues studied were so complicated, and took so many years, that it has been difficult to appreciate the magnitude of their accomplishments. It is therefore understandable why his work has had such a great impact on our lives. Almost single-handedly he created the fields of communication acoustics and speech and hearing as we know them today. Everyone who has ever used the telephone has reaped the benefit provided by this man and his genius. von Békésy, Davis, Stevens, and Zwicker are some of the names that come to mind when we think of hearing. Bell invented the telephone, and Edison made it into a practical device. Harvey Fletcher may not be as well known as these men today, but his scientific contributions to the fields of telephony, hearing, and human communication are absolutely unsurpassed. Given this present opportunity to reflect back on this great man, I would describe Harvey Fletcher as the singular intellectual force in the development of present-day communication acoustics and telephony.

  5. Multilevel Analysis in Analyzing Speech Data

    ERIC Educational Resources Information Center

    Guddattu, Vasudeva; Krishna, Y.

    2011-01-01

    The speech produced by human vocal tract is a complex acoustic signal, with diverse applications in phonetics, speech synthesis, automatic speech recognition, speaker identification, communication aids, speech pathology, speech perception, machine translation, hearing research, rehabilitation and assessment of communication disorders and many…

  6. Denoising of human speech using combined acoustic and em sensor signal processing

    SciTech Connect

    Ng, L C; Burnett, G C; Holzrichter, J F; Gable, T J

    1999-11-29

    Low Power EM radar-like sensors have made it possible to measure properties of the human speech production system in real-time, without acoustic interference. This greatly enhances the quality and quantify of information for many speech related applications. See Holzrichter, Burnett, Ng, and Lea, J. Acoustic. Soc. Am. 103 (1) 622 (1998). By using combined Glottal-EM- Sensor- and Acoustic-signals, segments of voiced, unvoiced, and no-speech can be reliably defined. Real-time Denoising filters can be constructed to remove noise from the user's corresponding speech signal.

  7. Virtual acoustics, aeronautics, and communications.

    PubMed

    Begault, D R

    1998-06-01

    An optimal approach to auditory display design for commercial aircraft would utilize both spatialized (3-D) audio techniques and active noise cancellation for safer operations. Results from several aircraft simulator studies conducted at NASA Ames Research Center are reviewed, including Traffic alert and Collision Avoidance System (TCAS) warnings, spoken orientation "beacons" for gate identification and collision avoidance on the ground, and hardware for improved speech intelligibility. The implications of hearing loss among pilots is also considered.

  8. Virtual acoustics, aeronautics, and communications

    NASA Technical Reports Server (NTRS)

    Begault, D. R.; Wenzel, E. M. (Principal Investigator)

    1998-01-01

    An optimal approach to auditory display design for commercial aircraft would utilize both spatialized (3-D) audio techniques and active noise cancellation for safer operations. Results from several aircraft simulator studies conducted at NASA Ames Research Center are reviewed, including Traffic alert and Collision Avoidance System (TCAS) warnings, spoken orientation "beacons" for gate identification and collision avoidance on the ground, and hardware for improved speech intelligibility. The implications of hearing loss among pilots is also considered.

  9. Virtual Acoustics, Aeronautics and Communications

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Null, Cynthia H. (Technical Monitor)

    1996-01-01

    An optimal approach to auditory display design for commercial aircraft would utilize both spatialized ("3-D") audio techniques and active noise cancellation for safer operations. Results from several aircraft simulator studies conducted at NASA Ames Research Center are reviewed, including Traffic alert and Collision Avoidance System (TCAS) warnings, spoken orientation "beacons" for gate identification and collision avoidance on the ground, and hardware for improved speech intelligibility. The implications of hearing loss amongst pilots is also considered.

  10. An acoustic comparison of two women's infant- and adult-directed speech

    NASA Astrophysics Data System (ADS)

    Andruski, Jean; Katz-Gershon, Shiri

    2003-04-01

    In addition to having prosodic characteristics that are attractive to infant listeners, infant-directed (ID) speech shares certain characteristics of adult-directed (AD) clear speech, such as increased acoustic distance between vowels, that might be expected to make ID speech easier for adults to perceive in noise than AD conversational speech. However, perceptual tests of two women's ID productions by Andruski and Bessega [J. Acoust. Soc. Am. 112, 2355] showed that is not always the case. In a word identification task that compared ID speech with AD clear and conversational speech, one speaker's ID productions were less well-identified than AD clear speech, but better identified than AD conversational speech. For the second woman, ID speech was the least accurately identified of the three speech registers. For both speakers, hard words (infrequent words with many lexical neighbors) were also at an increased disadvantage relative to easy words (frequent words with few lexical neighbors) in speech registers that were less accurately perceived. This study will compare several acoustic properties of these women's productions, including pitch and formant-frequency characteristics. Results of the acoustic analyses will be examined with the original perceptual results to suggest reasons for differences in listener's accuracy in identifying these two women's ID speech in noise.

  11. Acoustic Analysis of the Voiced-Voiceless Distinction in Dutch Tracheoesophageal Speech

    ERIC Educational Resources Information Center

    Jongmans, Petra; Wempe, Ton G.; van Tinteren, Harm; Hilgers, Frans J. M.; Pols, Louis C. W.; van As-Brooks, Corina J.

    2010-01-01

    Purpose: Confusions between voiced and voiceless plosives and voiced and voiceless fricatives are common in Dutch tracheoesophageal (TE) speech. This study investigates (a) which acoustic measures are found to convey a correct voicing contrast in TE speech and (b) whether different measures are found in TE speech than in normal laryngeal (NL)…

  12. Perceptual and Acoustic Reliability Estimates for the Speech Disorders Classification System (SDCS)

    ERIC Educational Resources Information Center

    Shriberg, Lawrence D.; Fourakis, Marios; Hall, Sheryl D.; Karlsson, Heather B.; Lohmeier, Heather L.; McSweeny, Jane L.; Potter, Nancy L.; Scheer-Cohen, Alison R.; Strand, Edythe A.; Tilkens, Christie M.; Wilson, David L.

    2010-01-01

    A companion paper describes three extensions to a classification system for paediatric speech sound disorders termed the Speech Disorders Classification System (SDCS). The SDCS uses perceptual and acoustic data reduction methods to obtain information on a speaker's speech, prosody, and voice. The present paper provides reliability estimates for…

  13. Teaching Speech Communication from a Critical Thinking Perspective.

    ERIC Educational Resources Information Center

    O'Neill, Daniel J.

    Teaching speech communication from a critical thinking perspective involves specific instructional approaches and strategies. And the perspective can be applied to "survey courses" such as communication theory, rhetorical theory, nonverbal communication, and organizational communication theory. When a student engages in an internal or…

  14. Acoustic communication in insect disease vectors

    PubMed Central

    Vigoder, Felipe de Mello; Ritchie, Michael Gordon; Gibson, Gabriella; Peixoto, Alexandre Afranio

    2013-01-01

    Acoustic signalling has been extensively studied in insect species, which has led to a better understanding of sexual communication, sexual selection and modes of speciation. The significance of acoustic signals for a blood-sucking insect was first reported in the XIX century by Christopher Johnston, studying the hearing organs of mosquitoes, but has received relatively little attention in other disease vectors until recently. Acoustic signals are often associated with mating behaviour and sexual selection and changes in signalling can lead to rapid evolutionary divergence and may ultimately contribute to the process of speciation. Songs can also have implications for the success of novel methods of disease control such as determining the mating competitiveness of modified insects used for mass-release control programs. Species-specific sound “signatures” may help identify incipient species within species complexes that may be of epidemiological significance, e.g. of higher vectorial capacity, thereby enabling the application of more focussed control measures to optimise the reduction of pathogen transmission. Although the study of acoustic communication in insect vectors has been relatively limited, this review of research demonstrates their value as models for understanding both the functional and evolutionary significance of acoustic communication in insects. PMID:24473800

  15. Moving to the Speed of Sound: Context Modulation of the Effect of Acoustic Properties of Speech

    ERIC Educational Resources Information Center

    Shintel, Hadas; Nusbaum, Howard C.

    2008-01-01

    Suprasegmental acoustic patterns in speech can convey meaningful information and affect listeners' interpretation in various ways, including through systematic analog mapping of message-relevant information onto prosody. We examined whether the effect of analog acoustic variation is governed by the acoustic properties themselves. For example, fast…

  16. Hate Speech and Its Harms: A Communication Theory Perspective.

    ERIC Educational Resources Information Center

    Calvert, Clay

    1997-01-01

    Uses J.W. Carey's transmission and ritual communication models to illustrate harms caused by hate speech. Examines legal precedent for the ritual model, but suggests that courts more often adopt the transmission model. Argues that, although the ritual model points to a different, long-term harm caused by hate speech, its adoption raises troubling…

  17. Mandarin Speech Perception in Combined Electric and Acoustic Stimulation

    PubMed Central

    Li, Yongxin; Zhang, Guoping; Galvin, John J.; Fu, Qian-Jie

    2014-01-01

    For deaf individuals with residual low-frequency acoustic hearing, combined use of a cochlear implant (CI) and hearing aid (HA) typically provides better speech understanding than with either device alone. Because of coarse spectral resolution, CIs do not provide fundamental frequency (F0) information that contributes to understanding of tonal languages such as Mandarin Chinese. The HA can provide good representation of F0 and, depending on the range of aided acoustic hearing, first and second formant (F1 and F2) information. In this study, Mandarin tone, vowel, and consonant recognition in quiet and noise was measured in 12 adult Mandarin-speaking bimodal listeners with the CI-only and with the CI+HA. Tone recognition was significantly better with the CI+HA in noise, but not in quiet. Vowel recognition was significantly better with the CI+HA in quiet, but not in noise. There was no significant difference in consonant recognition between the CI-only and the CI+HA in quiet or in noise. There was a wide range in bimodal benefit, with improvements often greater than 20 percentage points in some tests and conditions. The bimodal benefit was compared to CI subjects’ HA-aided pure-tone average (PTA) thresholds between 250 and 2000 Hz; subjects were divided into two groups: “better” PTA (<50 dB HL) or “poorer” PTA (>50 dB HL). The bimodal benefit differed significantly between groups only for consonant recognition. The bimodal benefit for tone recognition in quiet was significantly correlated with CI experience, suggesting that bimodal CI users learn to better combine low-frequency spectro-temporal information from acoustic hearing with temporal envelope information from electric hearing. Given the small number of subjects in this study (n = 12), further research with Chinese bimodal listeners may provide more information regarding the contribution of acoustic and electric hearing to tonal language perception. PMID:25386962

  18. Multipath correlations in underwater acoustic communication channels.

    PubMed

    Huang, S H; Yang, T C; Huang, Chen-Fen

    2013-04-01

    Uncorrelated scattering (US), which assumes that multipath arrivals undergo uncorrelated scattering and are thus uncorrelated, has been the standard model for digital communications including underwater acoustic communications. This paper examines the cross-correlation of multipath arrivals based on at-sea data with different temporal coherence time, assuming quasi-stationary statistics. It is found that multipath arrivals are highly cross-correlated when the channel is temporally coherent, and are uncorrelated when the channel is temporally incoherent. A theoretical model based on the path phase rates and relative-phase fluctuations is used to explain experimentally observed phenomena, assuming the path amplitudes vary slowly compared with the phases. The implications of correlated scattering for underwater acoustic communication channel tracking are discussed.

  19. Acoustical Characteristics of Mastication Sounds: Application of Speech Analysis Techniques

    NASA Astrophysics Data System (ADS)

    Brochetti, Denise

    Food scientists have used acoustical methods to study characteristics of mastication sounds in relation to food texture. However, a model for analysis of the sounds has not been identified, and reliability of the methods has not been reported. Therefore, speech analysis techniques were applied to mastication sounds, and variation in measures of the sounds was examined. To meet these objectives, two experiments were conducted. In the first experiment, a digital sound spectrograph generated waveforms and wideband spectrograms of sounds by 3 adult subjects (1 male, 2 females) for initial chews of food samples differing in hardness and fracturability. Acoustical characteristics were described and compared. For all sounds, formants appeared in the spectrograms, and energy occurred across a 0 to 8000-Hz range of frequencies. Bursts characterized waveforms for peanut, almond, raw carrot, ginger snap, and hard candy. Duration and amplitude of the sounds varied with the subjects. In the second experiment, the spectrograph was used to measure the duration, amplitude, and formants of sounds for the initial 2 chews of cylindrical food samples (raw carrot, teething toast) differing in diameter (1.27, 1.90, 2.54 cm). Six adult subjects (3 males, 3 females) having normal occlusions and temporomandibular joints chewed the samples between the molar teeth and with the mouth open. Ten repetitions per subject were examined for each food sample. Analysis of estimates of variation indicated an inconsistent intrasubject variation in the acoustical measures. Food type and sample diameter also affected the estimates, indicating the variable nature of mastication. Generally, intrasubject variation was greater than intersubject variation. Analysis of ranks of the data indicated that the effect of sample diameter on the acoustical measures was inconsistent and depended on the subject and type of food. If inferences are to be made concerning food texture from acoustical measures of mastication

  20. Rocking Boats, Blowing Whistles, and Teaching Speech Communication.

    ERIC Educational Resources Information Center

    Redding, W. Charles

    1985-01-01

    Examines and defends two forms of dissent in organizational life, "rocking boats" and "blowing whistles." Suggests ways speech communication teachers can prepare students to cope with the organizational world. (PD)

  1. Lest We Forget: The Role of Internships in Speech Communication.

    ERIC Educational Resources Information Center

    Porter, Sharon B.

    Two observations point to the need for improvement of teacher internship programs in speech communication: (1) the supply of education graduates is far greater than the demand, and (2) in many universities and colleges, communication departments are struggling for survival. Scholars of communication must provide the most competent leadership…

  2. Communication without Speech: A Guide for Parents and Teachers.

    ERIC Educational Resources Information Center

    Bloomberg, Karen, Ed.; Johnson, Hilary, Ed.

    This guide addresses issues facing the parents, teachers and caregivers of children who are unable to use normal speech as a means of communication. It focuses on people who are intellectually disabled or children who are starting to use augmentative communication. The guide includes the following topics: the nature of communication; an overview…

  3. The minor third communicates sadness in speech, mirroring its use in music.

    PubMed

    Curtis, Meagan E; Bharucha, Jamshed J

    2010-06-01

    There is a long history of attempts to explain why music is perceived as expressing emotion. The relationship between pitches serves as an important cue for conveying emotion in music. The musical interval referred to as the minor third is generally thought to convey sadness. We reveal that the minor third also occurs in the pitch contour of speech conveying sadness. Bisyllabic speech samples conveying four emotions were recorded by 9 actresses. Acoustic analyses revealed that the relationship between the 2 salient pitches of the sad speech samples tended to approximate a minor third. Participants rated the speech samples for perceived emotion, and the use of numerous acoustic parameters as cues for emotional identification was modeled using regression analysis. The minor third was the most reliable cue for identifying sadness. Additional participants rated musical intervals for emotion, and their ratings verified the historical association between the musical minor third and sadness. These findings support the theory that human vocal expressions and music share an acoustic code for communicating sadness.

  4. How stable are acoustic metrics of contrastive speech rhythm?

    PubMed

    Wiget, Lukas; White, Laurence; Schuppler, Barbara; Grenon, Izabelle; Rauch, Olesya; Mattys, Sven L

    2010-03-01

    Acoustic metrics of contrastive speech rhythm, based on vocalic and intervocalic interval durations, are intended to capture stable typological differences between languages. They should consequently be robust to variation between speakers, sentence materials, and measurers. This paper assesses the impact of these sources of variation on the metrics %V (proportion of utterance comprised of vocalic intervals), VarcoV (rate-normalized standard deviation of vocalic interval duration), and nPVI-V (a measure of the durational variability between successive pairs of vocalic intervals). Five measurers analyzed the same corpus of speech: five sentences read by six speakers of Standard Southern British English. Differences between sentences were responsible for the greatest variation in rhythm scores. Inter-speaker differences were also a source of significant variability. However, there was relatively little variation due to segmentation differences between measurers following an agreed protocol. An automated phone alignment process was also used: Rhythm scores thus derived showed good agreement with the human measurers. A number of recommendations for researchers wishing to exploit contrastive rhythm metrics are offered in conclusion. PMID:20329856

  5. Acoustic characteristics of Korean stops in Korean child-directed speech

    NASA Astrophysics Data System (ADS)

    Kim, Minjung; Stoel-Gammon, Carol

    2005-04-01

    A variety of cross-linguistic studies have documented that the acoustic properties of speech addressed to young children include exaggeration of pitch contours and acoustically salient features of phonetic units. It has been suggested that phonetic modifications of child-directed speech facilitate young children's speech perception by providing detailed phonetic information about the target word. While there are several studies reporting vowel modifications in speech to infants (i.e., hyper-articulated vowels), there has been relatively little research about consonant modifications in speech to young children (except for VOT). The present study examines acoustic properties of Korean stops in Korean mothers' speech to their children aged 29 to 38 months (N=6). Korean tense, lax, and aspirated stops are all voiceless in word-initial position, and are perceptually differentiated by several acoustic parameters including VOT, f0 of the following vowel, and the amplitude difference of the first and second harmonics at the voice onset of the following vowel. This study compares values of these parameters in Korean motherese to those in speech to adult Koreans from same speakers. Results focus on the acoustic properties of Korean stops in child-directed speech and how they are modified to help Korean young children learn the three-way phonetic contrast.

  6. The energetic basis of acoustic communication

    PubMed Central

    Gillooly, James F.; Ophir, Alexander G.

    2010-01-01

    Animals produce a tremendous diversity of sounds for communication to perform life's basic functions, from courtship and parental care to defence and foraging. Explaining this diversity in sound production is important for understanding the ecology, evolution and behaviour of species. Here, we present a theory of acoustic communication that shows that much of the heterogeneity in animal vocal signals can be explained based on the energetic constraints of sound production. The models presented here yield quantitative predictions on key features of acoustic signals, including the frequency, power and duration of signals. Predictions are supported with data from nearly 500 diverse species (e.g. insects, fishes, reptiles, amphibians, birds and mammals). These results indicate that, for all species, acoustic communication is primarily controlled by individual metabolism such that call features vary predictably with body size and temperature. These results also provide insights regarding the common energetic and neuromuscular constraints on sound production, and the ecological and evolutionary consequences of producing these sounds. PMID:20053641

  7. The energetic basis of acoustic communication.

    PubMed

    Gillooly, James F; Ophir, Alexander G

    2010-05-01

    Animals produce a tremendous diversity of sounds for communication to perform life's basic functions, from courtship and parental care to defence and foraging. Explaining this diversity in sound production is important for understanding the ecology, evolution and behaviour of species. Here, we present a theory of acoustic communication that shows that much of the heterogeneity in animal vocal signals can be explained based on the energetic constraints of sound production. The models presented here yield quantitative predictions on key features of acoustic signals, including the frequency, power and duration of signals. Predictions are supported with data from nearly 500 diverse species (e.g. insects, fishes, reptiles, amphibians, birds and mammals). These results indicate that, for all species, acoustic communication is primarily controlled by individual metabolism such that call features vary predictably with body size and temperature. These results also provide insights regarding the common energetic and neuromuscular constraints on sound production, and the ecological and evolutionary consequences of producing these sounds. PMID:20053641

  8. On the physics of underwater acoustic communications

    NASA Astrophysics Data System (ADS)

    Yang, T. C.

    2012-11-01

    Digital communications have been traditionally treated using signal processing approaches. Algorithm and performance is system-model dependent. For underwater acoustic communications, the system model is very complex due to the different channel environmental conditions which can vary rapidly with time, and are location dependent. Performance modeling is difficult without a profound channel model. Work over the last decades have shown that robust performance can be achieved by exploiting the channel physics allowing at the same time an initial estimation of the communication performance. This paper will review the basic signal processing approaches using languages familiar to the physicists/acousticians, point out the communication related physics issues, and discuss how channel physics can be exploited to improve the performance.

  9. Acoustic properties of naturally produced clear speech at normal speaking rates

    NASA Astrophysics Data System (ADS)

    Krause, Jean C.; Braida, Louis D.

    2004-01-01

    Sentences spoken ``clearly'' are significantly more intelligible than those spoken ``conversationally'' for hearing-impaired listeners in a variety of backgrounds [Picheny et al., J. Speech Hear. Res. 28, 96-103 (1985); Uchanski et al., ibid. 39, 494-509 (1996); Payton et al., J. Acoust. Soc. Am. 95, 1581-1592 (1994)]. While producing clear speech, however, talkers often reduce their speaking rate significantly [Picheny et al., J. Speech Hear. Res. 29, 434-446 (1986); Uchanski et al., ibid. 39, 494-509 (1996)]. Yet speaking slowly is not solely responsible for the intelligibility benefit of clear speech (over conversational speech), since a recent study [Krause and Braida, J. Acoust. Soc. Am. 112, 2165-2172 (2002)] showed that talkers can produce clear speech at normal rates with training. This finding suggests that clear speech has inherent acoustic properties, independent of rate, that contribute to improved intelligibility. Identifying these acoustic properties could lead to improved signal processing schemes for hearing aids. To gain insight into these acoustical properties, conversational and clear speech produced at normal speaking rates were analyzed at three levels of detail (global, phonological, and phonetic). Although results suggest that talkers may have employed different strategies to achieve clear speech at normal rates, two global-level properties were identified that appear likely to be linked to the improvements in intelligibility provided by clear/normal speech: increased energy in the 1000-3000-Hz range of long-term spectra and increased modulation depth of low frequency modulations of the intensity envelope. Other phonological and phonetic differences associated with clear/normal speech include changes in (1) frequency of stop burst releases, (2) VOT of word-initial voiceless stop consonants, and (3) short-term vowel spectra.

  10. DARPA TIMIT acoustic-phonetic continous speech corpus CD-ROM. NIST speech disc 1-1.1

    NASA Astrophysics Data System (ADS)

    Garofolo, J. S.; Lamel, L. F.; Fisher, W. M.; Fiscus, J. G.; Pallett, D. S.

    1993-02-01

    The Texas Instruments/Massachusetts Institute of Technology (TIMIT) corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems. TIMIT contains speech from 630 speakers representing 8 major dialect divisions of American English, each speaking 10 phonetically-rich sentences. The TIMIT corpus includes time-aligned orthographic, phonetic, and word transcriptions, as well as speech waveform data for each spoken sentence. The release of TIMIT contains several improvements over the Prototype CD-ROM released in December, 1988: (1) full 630-speaker corpus, (2) checked and corrected transcriptions, (3) word-alignment transcriptions, (4) NIST SPHERE-headered waveform files and header manipulation software, (5) phonemic dictionary, (6) new test and training subsets balanced for dialectal and phonetic coverage, and (7) more extensive documentation.

  11. Formant Centralization Ratio: A Proposal for a New Acoustic Measure of Dysarthric Speech

    ERIC Educational Resources Information Center

    Sapir, Shimon; Ramig, Lorraine O.; Spielman, Jennifer L.; Fox, Cynthia

    2010-01-01

    Purpose: The vowel space area (VSA) has been used as an acoustic metric of dysarthric speech, but with varying degrees of success. In this study, the authors aimed to test an alternative metric to the VSA--the "formant centralization ratio" (FCR), which is hypothesized to more effectively differentiate dysarthric from healthy speech and register…

  12. Effects of Training on the Acoustic-Phonetic Representation of Synthetic Speech

    ERIC Educational Resources Information Center

    Francis, Alexander L.; Nusbaum, Howard C.; Fenn, Kimberly

    2007-01-01

    Purpose: Investigate training-related changes in acoustic-phonetic representation of consonants produced by a text-to-speech (TTS) computer speech synthesizer. Method: Forty-eight adult listeners were trained to better recognize words produced by a TTS system. Nine additional untrained participants served as controls. Before and after training,…

  13. General Education Standards for Speech Communication in Maryland Higher Education.

    ERIC Educational Resources Information Center

    Engleberg, Isa N.

    On September 26, 1998, the Maryland Communication Association unanimously approved the "Standards for General Education Speech Communication Courses in Maryland Higher Education Institutions." On February 10, 1999, the state's intersegmental Chief Academic Officers (CAO) group endorsed the standards for inclusion in their "Guiding Principles for…

  14. Bibliographic Annual in Speech Communication: 1974. Volume 5.

    ERIC Educational Resources Information Center

    Kennicott, Patrick C., Ed.

    This annotated bibliography is an annual volume devoted to maintaining a record of graduate work in speech communication, providing abstracts of doctoral dissertations, and making subject area bibliographies available. The contents of this volume include "Studies in Mass Communication: A Selected Bibliography, 1973" by Roland C. Johnson and…

  15. The psychology of songbird acoustic communication: Open-ended categorization and its role in songbird perception

    NASA Astrophysics Data System (ADS)

    Sturdy, Christopher

    2005-09-01

    Songbirds are used as a model system for understanding vocal learning, production, and perception. Over the past several years one particular aspect of songbird communication has been examined, vocal perception that has parallels with human speech perception, namely, the perception of natural acoustic categories in songbird vocalizations. A series of studies will be presented highlighting the role of open-ended categorization in the perception of vocalizations in songbirds. Attempts will be made to make the case that open-ended categorization is a fundamental proximate mechanism for acoustic communication in both humans and songbirds, that it is used in a highly analogous manner in each, and future directions will attempt to integrate these two areas of research and, ultimately, lead to a more complete understanding of acoustic communication in animals.

  16. Existence detection and embedding rate estimation of blended speech in covert speech communications.

    PubMed

    Li, Lijuan; Gao, Yong

    2016-01-01

    Covert speech communications may be used by terrorists to commit crimes through Internet. Steganalysis aims to detect secret information in covert communications to prevent crimes. Herein, based on the average zero crossing rate of the odd-even difference (AZCR-OED), a steganalysis algorithm for blended speech is proposed; it can detect the existence and estimate the embedding rate of blended speech. First, the odd-even difference (OED) of the speech signal is calculated and divided into frames. The average zero crossing rate (ZCR) is calculated for each OED frame, and the minimum average ZCR and AZCR-OED of the entire speech signal are extracted as features. Then, a support vector machine classifier is used to determine whether the speech signal is blended. Finally, a voice activity detection algorithm is applied to determine the hidden location of the secret speech and estimate the embedding rate. The results demonstrate that without attack, the detection accuracy can reach 80 % or more when the embedding rate is greater than 10 %, and the estimated embedding rate is similar to the real value. And when some attacks occur, it can also reach relatively high detection accuracy. The algorithm has high performance in terms of accuracy, effectiveness and robustness. PMID:27462497

  17. Cognitive Spare Capacity and Speech Communication: A Narrative Overview

    PubMed Central

    2014-01-01

    Background noise can make speech communication tiring and cognitively taxing, especially for individuals with hearing impairment. It is now well established that better working memory capacity is associated with better ability to understand speech under adverse conditions as well as better ability to benefit from the advanced signal processing in modern hearing aids. Recent work has shown that although such processing cannot overcome hearing handicap, it can increase cognitive spare capacity, that is, the ability to engage in higher level processing of speech. This paper surveys recent work on cognitive spare capacity and suggests new avenues of investigation. PMID:24971355

  18. The Use of Artificial Neural Networks to Estimate Speech Intelligibility from Acoustic Variables: A Preliminary Analysis.

    ERIC Educational Resources Information Center

    Metz, Dale Evan; And Others

    1992-01-01

    A preliminary scheme for estimating the speech intelligibility of hearing-impaired speakers from acoustic parameters, using a computerized artificial neural network to process mathematically the acoustic input variables, is outlined. Tests with 60 hearing-impaired speakers found the scheme to be highly accurate in identifying speakers separated by…

  19. School cafeteria noise-The impact of room acoustics and speech intelligibility on children's voice levels

    NASA Astrophysics Data System (ADS)

    Bridger, Joseph F.

    2002-05-01

    The impact of room acoustics and speech intelligibility conditions of different school cafeterias on the voice levels of children is examined. Methods of evaluating cafeteria designs and predicting noise levels are discussed. Children are shown to modify their voice levels with changes in speech intelligibility like adults. Reverberation and signal to noise ratio are the important acoustical factors affecting speech intelligibility. Children have much more difficulty than adults in conditions where noise and reverberation are present. To evaluate the relationship of voice level and speech intelligibility, a database of real sound levels and room acoustics data was generated from measurements and data recorded during visits to a variety of existing cafeterias under different occupancy conditions. The effects of speech intelligibility and room acoustics on childrens voice levels are demonstrated. A new method is presented for predicting speech intelligibility conditions and resulting noise levels for the design of new cafeterias and renovation of existing facilities. Measurements are provided for an existing school cafeteria before and after new room acoustics treatments were added. This will be helpful for acousticians, architects, school systems, regulatory agencies, and Parent Teacher Associations to create less noisy cafeteria environments.

  20. A maximum likelihood approach to estimating articulator positions from speech acoustics

    SciTech Connect

    Hogden, J.

    1996-09-23

    This proposal presents an algorithm called maximum likelihood continuity mapping (MALCOM) which recovers the positions of the tongue, jaw, lips, and other speech articulators from measurements of the sound-pressure waveform of speech. MALCOM differs from other techniques for recovering articulator positions from speech in three critical respects: it does not require training on measured or modeled articulator positions, it does not rely on any particular model of sound propagation through the vocal tract, and it recovers a mapping from acoustics to articulator positions that is linearly, not topographically, related to the actual mapping from acoustics to articulation. The approach categorizes short-time windows of speech into a finite number of sound types, and assumes the probability of using any articulator position to produce a given sound type can be described by a parameterized probability density function. MALCOM then uses maximum likelihood estimation techniques to: (1) find the most likely smooth articulator path given a speech sample and a set of distribution functions (one distribution function for each sound type), and (2) change the parameters of the distribution functions to better account for the data. Using this technique improves the accuracy of articulator position estimates compared to continuity mapping -- the only other technique that learns the relationship between acoustics and articulation solely from acoustics. The technique has potential application to computer speech recognition, speech synthesis and coding, teaching the hearing impaired to speak, improving foreign language instruction, and teaching dyslexics to read. 34 refs., 7 figs.

  1. Language-specific developmental differences in speech production: a cross-language acoustic study.

    PubMed

    Li, Fangfang

    2012-01-01

    Speech productions of 40 English- and 40 Japanese-speaking children (aged 2-5) were examined and compared with the speech produced by 20 adult speakers (10 speakers per language). Participants were recorded while repeating words that began with "s" and "sh" sounds. Clear language-specific patterns in adults' speech were found, with English speakers differentiating "s" and "sh" in 1 acoustic dimension (i.e., spectral mean) and Japanese speakers differentiating the 2 categories in 3 acoustic dimensions (i.e., spectral mean, standard deviation, and onset F2 frequency). For both language groups, children's speech exhibited a gradual change from an early undifferentiated form to later differentiated categories. The separation processes, however, only occur in those acoustic dimensions used by adults in the corresponding languages.

  2. Language Comprehension in Language-Learning Impaired Children Improved with Acoustically Modified Speech

    NASA Astrophysics Data System (ADS)

    Tallal, Paula; Miller, Steve L.; Bedi, Gail; Byma, Gary; Wang, Xiaoqin; Nagarajan, Srikantan S.; Schreiner, Christoph; Jenkins, William M.; Merzenich, Michael M.

    1996-01-01

    A speech processing algorithm was developed to create more salient versions of the rapidly changing elements in the acoustic waveform of speech that have been shown to be deficiently processed by language-learning impaired (LLI) children. LLI children received extensive daily training, over a 4-week period, with listening exercises in which all speech was translated into this synthetic form. They also received daily training with computer "games" designed to adaptively drive improvements in temporal processing thresholds. Significant improvements in speech discrimination and language comprehension abilities were demonstrated in two independent groups of LLI children.

  3. Costing Children's Speech, Language and Communication Interventions

    ERIC Educational Resources Information Center

    Beecham, Jennifer; Law, James; Zeng, Biao; Lindsay, Geoff

    2012-01-01

    Background: There are few economic evaluations of speech and language interventions. Such work requires underpinning by an accurate estimate of the costs of the intervention. This study seeks to address some of the complexities of this task by applying existing approaches of cost estimation to interventions described in published effectiveness…

  4. A Hybrid Acoustic and Pronunciation Model Adaptation Approach for Non-native Speech Recognition

    NASA Astrophysics Data System (ADS)

    Oh, Yoo Rhee; Kim, Hong Kook

    In this paper, we propose a hybrid model adaptation approach in which pronunciation and acoustic models are adapted by incorporating the pronunciation and acoustic variabilities of non-native speech in order to improve the performance of non-native automatic speech recognition (ASR). Specifically, the proposed hybrid model adaptation can be performed at either the state-tying or triphone-modeling level, depending at which acoustic model adaptation is performed. In both methods, we first analyze the pronunciation variant rules of non-native speakers and then classify each rule as either a pronunciation variant or an acoustic variant. The state-tying level hybrid method then adapts pronunciation models and acoustic models by accommodating the pronunciation variants in the pronunciation dictionary and by clustering the states of triphone acoustic models using the acoustic variants, respectively. On the other hand, the triphone-modeling level hybrid method initially adapts pronunciation models in the same way as in the state-tying level hybrid method; however, for the acoustic model adaptation, the triphone acoustic models are then re-estimated based on the adapted pronunciation models and the states of the re-estimated triphone acoustic models are clustered using the acoustic variants. From the Korean-spoken English speech recognition experiments, it is shown that ASR systems employing the state-tying and triphone-modeling level adaptation methods can relatively reduce the average word error rates (WERs) by 17.1% and 22.1% for non-native speech, respectively, when compared to a baseline ASR system.

  5. The acoustics for speech of eight auditoriums in the city of Sao Paulo

    NASA Astrophysics Data System (ADS)

    Bistafa, Sylvio R.

    2002-11-01

    Eight auditoriums with a proscenium type of stage, which usually operate as dramatic theaters in the city of Sao Paulo, were acoustically surveyed in terms of their adequacy to unassisted speech. Reverberation times, early decay times, and speech levels were measured in different positions, together with objective measures of speech intelligibility. The measurements revealed reverberation time values rather uniform throughout the rooms, whereas significant variations were found in the values of the other acoustical measures with position. The early decay time was found to be better correlated with the objective measures of speech intelligibility than the reverberation time. The results from the objective measurements of speech intelligibility revealed that the speech transmission index STI, and its simplified version RaSTI, are strongly correlated with the early-to-late sound ratio C50 (1 kHz). However, it was found that the criterion value of acceptability of the latter is more easily met than the former. The results from these measurements enable to understand how the characteristics of the architectural design determine the acoustical quality for speech. Measurements of ST1-Gade were made as an attempt to validate it as an objective measure of ''support'' for the actor. The preliminary diagnosing results with ray tracing simulations will also be presented.

  6. A magnetic resonance imaging study on the articulatory and acoustic speech parameters of Malay vowels

    PubMed Central

    2014-01-01

    The phonetic properties of six Malay vowels are investigated using magnetic resonance imaging (MRI) to visualize the vocal tract in order to obtain dynamic articulatory parameters during speech production. To resolve image blurring due to the tongue movement during the scanning process, a method based on active contour extraction is used to track tongue contours. The proposed method efficiently tracks tongue contours despite the partial blurring of MRI images. Consequently, the articulatory parameters that are effectively measured as tongue movement is observed, and the specific shape of the tongue and its position for all six uttered Malay vowels are determined. Speech rehabilitation procedure demands some kind of visual perceivable prototype of speech articulation. To investigate the validity of the measured articulatory parameters based on acoustic theory of speech production, an acoustic analysis based on the uttered vowels by subjects has been performed. As the acoustic speech and articulatory parameters of uttered speech were examined, a correlation between formant frequencies and articulatory parameters was observed. The experiments reported a positive correlation between the constriction location of the tongue body and the first formant frequency, as well as a negative correlation between the constriction location of the tongue tip and the second formant frequency. The results demonstrate that the proposed method is an effective tool for the dynamic study of speech production. PMID:25060583

  7. Speech and melody recognition in binaurally combined acoustic and electric hearing

    NASA Astrophysics Data System (ADS)

    Kong, Ying-Yee; Stickney, Ginger S.; Zeng, Fan-Gang

    2005-03-01

    Speech recognition in noise and music perception is especially challenging for current cochlear implant users. The present study utilizes the residual acoustic hearing in the nonimplanted ear in five cochlear implant users to elucidate the role of temporal fine structure at low frequencies in auditory perception and to test the hypothesis that combined acoustic and electric hearing produces better performance than either mode alone. The first experiment measured speech recognition in the presence of competing noise. It was found that, although the residual low-frequency (<1000 Hz) acoustic hearing produced essentially no recognition for speech recognition in noise, it significantly enhanced performance when combined with the electric hearing. The second experiment measured melody recognition in the same group of subjects and found that, contrary to the speech recognition result, the low-frequency acoustic hearing produced significantly better performance than the electric hearing. It is hypothesized that listeners with combined acoustic and electric hearing might use the correlation between the salient pitch in low-frequency acoustic hearing and the weak pitch in the envelope to enhance segregation between signal and noise. The present study suggests the importance and urgency of accurately encoding the fine-structure cue in cochlear implants. .

  8. Behavioral Objectives for Speech-Communication: Panacea, Pandemonium, Or . . . .

    ERIC Educational Resources Information Center

    Parks, Arlie Muller

    The use of behavioral objectives is challenged on three bases: feasibility, utility, and advisability. Behavioral objectives, however, are useful in teaching speech communication. They allow the teacher to start to pinpoint the items the student needs to practice in order to improve. If students are trained to decide which behaviors they wish to…

  9. Business Communication Students Learn to Hear a Bad Speech Habit

    ERIC Educational Resources Information Center

    Bell, Reginald L.; Liang-Bell, Lei Paula; Deselle, Bettye

    2006-01-01

    Students were trained to perceive filled pauses (FP) as a bad speech habit. In a series of classroom sensitivity training activities, followed by students being rewarded to observe twenty minutes of live television from the public media, no differences between male and female Business Communication students was revealed. The practice of teaching…

  10. Adapting to the Job Market: Graduate Programs in Speech Communication.

    ERIC Educational Resources Information Center

    Berg, David M.

    The percentage of speech communication doctoral graduates employed full time and the percentage working in academic institutions have declined considerably since 1968. The glut of humanities doctorates appears to present three courses of action: increase undergraduate enrollments, decrease graduate enrollments, or increase nonacademic employment…

  11. Communication in a noisy environment: Perception of one's own voice and speech enhancement

    NASA Astrophysics Data System (ADS)

    Le Cocq, Cecile

    Workers in noisy industrial environments are often confronted to communication problems. Lost of workers complain about not being able to communicate easily with their coworkers when they wear hearing protectors. In consequence, they tend to remove their protectors, which expose them to the risk of hearing loss. In fact this communication problem is a double one: first the hearing protectors modify one's own voice perception; second they interfere with understanding speech from others. This double problem is examined in this thesis. When wearing hearing protectors, the modification of one's own voice perception is partly due to the occlusion effect which is produced when an earplug is inserted in the car canal. This occlusion effect has two main consequences: first the physiological noises in low frequencies are better perceived, second the perception of one's own voice is modified. In order to have a better understanding of this phenomenon, the literature results are analyzed systematically, and a new method to quantify the occlusion effect is developed. Instead of stimulating the skull with a bone vibrator or asking the subject to speak as is usually done in the literature, it has been decided to excite the buccal cavity with an acoustic wave. The experiment has been designed in such a way that the acoustic wave which excites the buccal cavity does not excite the external car or the rest of the body directly. The measurement of the hearing threshold in open and occluded car has been used to quantify the subjective occlusion effect for an acoustic wave in the buccal cavity. These experimental results as well as those reported in the literature have lead to a better understanding of the occlusion effect and an evaluation of the role of each internal path from the acoustic source to the internal car. The speech intelligibility from others is altered by both the high sound levels of noisy industrial environments and the speech signal attenuation due to hearing

  12. Is Birdsong More Like Speech or Music?

    PubMed

    Shannon, Robert V

    2016-04-01

    Music and speech share many acoustic cues but not all are equally important. For example, harmonic pitch is essential for music but not for speech. When birds communicate is their song more like speech or music? A new study contrasting pitch and spectral patterns shows that birds perceive their song more like humans perceive speech. PMID:26944220

  13. Mathematical model of acoustic speech production with mobile walls of the vocal tract

    NASA Astrophysics Data System (ADS)

    Lyubimov, N. A.; Zakharov, E. V.

    2016-03-01

    A mathematical speech production model is considered that describes acoustic oscillation propagation in a vocal tract with mobile walls. The wave field function satisfies the Helmholtz equation with boundary conditions of the third kind (impedance type). The impedance mode corresponds to a threeparameter pendulum oscillation model. The experimental research demonstrates the nonlinear character of how the mobility of the vocal tract walls influence the spectral envelope of a speech signal.

  14. Collaboration between Teachers and Speech and Language Therapists: Services for Primary School Children with Speech, Language and Communication Needs

    ERIC Educational Resources Information Center

    Glover, Anna; McCormack, Jane; Smith-Tamaray, Michelle

    2015-01-01

    Speech, language and communication needs (SLCN) are prevalent among primary school-aged children. Collaboration between speech and language therapists (SLTs) and teachers is beneficial for supporting children's communication skills. The aim of this study was to investigate the needs of both professional groups and their preferences for service…

  15. Vowel Acoustics in Adults with Apraxia of Speech

    ERIC Educational Resources Information Center

    Jacks, Adam; Mathes, Katey A.; Marquardt, Thomas P.

    2010-01-01

    Purpose: To investigate the hypothesis that vowel production is more variable in adults with acquired apraxia of speech (AOS) relative to healthy individuals with unimpaired speech. Vowel formant frequency measures were selected as the specific target of focus. Method: Seven adults with AOS and aphasia produced 15 repetitions of 6 American English…

  16. Precategorical Acoustic Storage and the Perception of Speech

    ERIC Educational Resources Information Center

    Frankish, Clive

    2008-01-01

    Theoretical accounts of both speech perception and of short term memory must consider the extent to which perceptual representations of speech sounds might survive in relatively unprocessed form. This paper describes a novel version of the serial recall task that can be used to explore this area of shared interest. In immediate recall of digit…

  17. Acoustic properties of vowels in clear and conversational speech by female non-native English speakers

    NASA Astrophysics Data System (ADS)

    Li, Chi-Nin; So, Connie K.

    2005-04-01

    Studies have shown that talkers can improve the intelligibility of their speech when instructed to speak as if talking to a hearing-impaired person. The improvement of speech intelligibility is associated with specific acoustic-phonetic changes: increases in vowel duration and fundamental frequency (F0), a wider pitch range, and a shift in formant frequencies for F1 and F2. Most previous studies of clear speech production have been conducted with native speakers; research with second language speakers is much less common. The present study examined the acoustic properties of non-native English vowels produced in a clear speaking style. Five female Cantonese speakers and a comparison group of English speakers were recorded producing four vowels (/i u ae a/) in /bVt/ context in conversational and clear speech. Vowel durations, F0, pitch range, and the first two formants for each of the four vowels were measured. Analyses revealed that for both groups of speakers, vowel durations, F0, pitch range, and F1 spoken clearly were greater than those produced conversationally. However, F2 was higher in conversational speech than in clear speech. The findings suggest that female non-native English speakers exhibit acoustic-phonetic patterns similar to those of native speakers when asked to produce English vowels clearly.

  18. Speech privacy and annoyance considerations in the acoustic environment of passenger cars of high-speed trains.

    PubMed

    Jeon, Jin Yong; Hong, Joo Young; Jang, Hyung Suk; Kim, Jae Hyeon

    2015-12-01

    It is necessary to consider not only annoyance of interior noises but also speech privacy to achieve acoustic comfort in a passenger car of a high-speed train because speech from other passengers can be annoying. This study aimed to explore an optimal acoustic environment to satisfy speech privacy and reduce annoyance in a passenger car. Two experiments were conducted using speech sources and compartment noise of a high speed train with varying speech-to-noise ratios (SNRA) and background noise levels (BNL). Speech intelligibility was tested in experiment I, and in experiment II, perceived speech privacy, annoyance, and acoustic comfort of combined sounds with speech and background noise were assessed. The results show that speech privacy and annoyance were significantly influenced by the SNRA. In particular, the acoustic comfort was evaluated as acceptable when the SNRA was less than -6 dB for both speech privacy and noise annoyance. In addition, annoyance increased significantly as the BNL exceeded 63 dBA, whereas the effect of the background-noise level on the speech privacy was not significant. These findings suggest that an optimal level of interior noise in a passenger car might exist between 59 and 63 dBA, taking normal speech levels into account.

  19. Education in acoustics and speech science using vocal-tract models.

    PubMed

    Arai, Takayuki

    2012-03-01

    Several vocal-tract models were reviewed, with special focus given to the sliding vocal-tract model [T. Arai, Acoust. Sci. Technol. 27(6), 384-388 (2006)]. All of the models have been shown to be excellent tools for teaching acoustics and speech science to elementary through university level students. The sliding three-tube model is based on Fant's three-tube model [G. Fant, Acoustic Theory of Speech Production (Mouton, The Hague, The Netherlands, 2006)] and consists of a long tube with a slider simulating tongue constriction. In this article, the design of the sliding vocal-tract model was reviewed. Then a science workshop was discussed where children were asked to make their own sliding vocal-tract models using simple materials. It was also discussed how the sliding vocal-tract model compares to our other vocal-tract models, emphasizing how the model can be used to instruct students at higher levels, such as undergraduate and graduate education in acoustics and speech science. Through this discussion the vocal-tract models were shown to be a powerful tool for education in acoustics and speech science for all ages of students.

  20. Augmentative and Alternative Communication in Autism: A Comparison of the Picture Exchange Communication System and Speech-Output Technology

    ERIC Educational Resources Information Center

    Boesch, Miriam Chacon

    2011-01-01

    The purpose of this comparative efficacy study was to investigate the Picture Exchange Communication System (PECS) and a speech-generating device (SGD) in developing requesting skills, social-communicative behavior, and speech for three elementary-age children with severe autism and little to no functional speech. Requesting was selected as the…

  1. Speech coding at 4800 bps for mobile satellite communications

    NASA Technical Reports Server (NTRS)

    Gersho, Allen; Chan, Wai-Yip; Davidson, Grant; Chen, Juin-Hwey; Yong, Mei

    1988-01-01

    A speech compression project has recently been completed to develop a speech coding algorithm suitable for operation in a mobile satellite environment aimed at providing telephone quality natural speech at 4.8 kbps. The work has resulted in two alternative techniques which achieve reasonably good communications quality at 4.8 kbps while tolerating vehicle noise and rather severe channel impairments. The algorithms are embodied in a compact self-contained prototype consisting of two AT and T 32-bit floating-point DSP32 digital signal processors (DSP). A Motorola 68HC11 microcomputer chip serves as the board controller and interface handler. On a wirewrapped card, the prototype's circuit footprint amounts to only 200 sq cm, and consumes about 9 watts of power.

  2. Towards personalized speech synthesis for augmentative and alternative communication.

    PubMed

    Mills, Timothy; Bunnell, H Timothy; Patel, Rupal

    2014-09-01

    Text-to-speech options on augmentative and alternative communication (AAC) devices are limited. Often, several individuals in a group setting use the same synthetic voice. This lack of customization may limit technology adoption and social integration. This paper describes our efforts to generate personalized synthesis for users with profoundly limited speech motor control. Existing voice banking and voice conversion techniques rely on recordings of clearly articulated speech from the target talker, which cannot be obtained from this population. Our VocaliD approach extracts prosodic properties from the target talker's source function and applies these features to a surrogate talker's database, generating a synthetic voice with the vocal identity of the target talker and the clarity of the surrogate talker. Promising intelligibility results suggest areas of further development for improved personalization. PMID:25025818

  3. Hello World, It's Me: Bringing the Basic Speech Communication Course into the Digital Age

    ERIC Educational Resources Information Center

    Kirkwood, Jessica; Gutgold, Nichola D.; Manley, Destiny

    2011-01-01

    During the past decade, instructors of speech communication have been adapting the introductory speech course to keep up with the television age. Learning units in speech textbooks now teach how to speak well on television, as well as how to interpret speeches in the media. This article argues that the computer age invites adaptation of the…

  4. Speech Acts across Cultures: Challenges to Communication in a Second Language. Studies on Language Acquisition, 11.

    ERIC Educational Resources Information Center

    Gass, Susan M., Ed.; Neu, Joyce, Ed.

    Articles on speech acts and intercultural communication include: "Investigating the Production of Speech Act Sets" (Andrew Cohen); "Non-Native Refusals: A Methodological Perspective" (Noel Houck, Susan M. Gass); "Natural Speech Act Data versus Written Questionnaire Data: How Data Collection Method Affects Speech Act Performance" (Leslie M. Beebe,…

  5. Perceiving speech in context: Compensation for contextual variability during acoustic cue encoding and categorization

    NASA Astrophysics Data System (ADS)

    Toscano, Joseph Christopher

    Several fundamental questions about speech perception concern how listeners understand spoken language despite considerable variability in speech sounds across different contexts (the problem of lack of invariance in speech). This contextual variability is caused by several factors, including differences between individual talkers' voices, variation in speaking rate, and effects of coarticulatory context. A number of models have been proposed to describe how the speech system handles differences across contexts. Critically, these models make different predictions about (1) whether contextual variability is handled at the level of acoustic cue encoding or categorization, (2) whether it is driven by feedback from category-level processes or interactions between cues, and (3) whether listeners discard fine-grained acoustic information to compensate for contextual variability. Separating the effects of cue- and category-level processing has been difficult because behavioral measures tap processes that occur well after initial cue encoding and are influenced by task demands and linguistic information. Recently, we have used the event-related brain potential (ERP) technique to examine cue encoding and online categorization. Specifically, we have looked at differences in the auditory N1 as a measure of acoustic cue encoding and the P3 as a measure of categorization. This allows us to examine multiple levels of processing during speech perception and can provide a useful tool for studying effects of contextual variability. Here, I apply this approach to determine the point in processing at which context has an effect on speech perception and to examine whether acoustic cues are encoded continuously. Several types of contextual variability (talker gender, speaking rate, and coarticulation), as well as several acoustic cues (voice onset time, formant frequencies, and bandwidths), are examined in a series of experiments. The results suggest that (1) at early stages of speech

  6. A speech processing study using an acoustic model of a multiple-channel cochlear implant

    NASA Astrophysics Data System (ADS)

    Xu, Ying

    1998-10-01

    A cochlear implant is an electronic device designed to provide sound information for adults and children who have bilateral profound hearing loss. The task of representing speech signals as electrical stimuli is central to the design and performance of cochlear implants. Studies have shown that the current speech- processing strategies provide significant benefits to cochlear implant users. However, the evaluation and development of speech-processing strategies have been complicated by hardware limitations and large variability in user performance. To alleviate these problems, an acoustic model of a cochlear implant with the SPEAK strategy is implemented in this study, in which a set of acoustic stimuli whose psychophysical characteristics are as close as possible to those produced by a cochlear implant are presented on normal-hearing subjects. To test the effectiveness and feasibility of this acoustic model, a psychophysical experiment was conducted to match the performance of a normal-hearing listener using model- processed signals to that of a cochlear implant user. Good agreement was found between an implanted patient and an age-matched normal-hearing subject in a dynamic signal discrimination experiment, indicating that this acoustic model is a reasonably good approximation of a cochlear implant with the SPEAK strategy. The acoustic model was then used to examine the potential of the SPEAK strategy in terms of its temporal and frequency encoding of speech. It was hypothesized that better temporal and frequency encoding of speech can be accomplished by higher stimulation rates and a larger number of activated channels. Vowel and consonant recognition tests were conducted on normal-hearing subjects using speech tokens processed by the acoustic model, with different combinations of stimulation rate and number of activated channels. The results showed that vowel recognition was best at 600 pps and 8 activated channels, but further increases in stimulation rate and

  7. Aging Affects Neural Synchronization to Speech-Related Acoustic Modulations

    PubMed Central

    Goossens, Tine; Vercammen, Charlotte; Wouters, Jan; van Wieringen, Astrid

    2016-01-01

    As people age, speech perception problems become highly prevalent, especially in noisy situations. In addition to peripheral hearing and cognition, temporal processing plays a key role in speech perception. Temporal processing of speech features is mediated by synchronized activity of neural oscillations in the central auditory system. Previous studies indicate that both the degree and hemispheric lateralization of synchronized neural activity relate to speech perception performance. Based on these results, we hypothesize that impaired speech perception in older persons may, in part, originate from deviances in neural synchronization. In this study, auditory steady-state responses that reflect synchronized activity of theta, beta, low and high gamma oscillations (i.e., 4, 20, 40, and 80 Hz ASSR, respectively) were recorded in young, middle-aged, and older persons. As all participants had normal audiometric thresholds and were screened for (mild) cognitive impairment, differences in synchronized neural activity across the three age groups were likely to be attributed to age. Our data yield novel findings regarding theta and high gamma oscillations in the aging auditory system. At an older age, synchronized activity of theta oscillations is increased, whereas high gamma synchronization is decreased. In contrast to young persons who exhibit a right hemispheric dominance for processing of high gamma range modulations, older adults show a symmetrical processing pattern. These age-related changes in neural synchronization may very well underlie the speech perception problems in aging persons. PMID:27378906

  8. Low-frequency speech cues and simulated electric-acoustic hearing

    PubMed Central

    Brown, Christopher A.; Bacon, Sid P.

    2009-01-01

    The addition of low-frequency acoustic information to real or simulated electric stimulation (so-called electric-acoustic stimulation or EAS) often results in large improvements in intelligibility, particularly in competing backgrounds. This may reflect the availability of fundamental frequency (F0) information in the acoustic region. The contributions of F0 and the amplitude envelope (as well as voicing) of speech to simulated EAS was examined by replacing the low-frequency speech with a tone that was modulated in frequency to track the F0 of the speech, in amplitude with the envelope of the low-frequency speech, or both. A four-channel vocoder simulated electric hearing. Significant benefit over vocoder alone was observed with the addition of a tone carrying F0 or envelope cues, and both cues combined typically provided significantly more benefit than either alone. The intelligibility improvement over vocoder was between 24 and 57 percentage points, and was unaffected by the presence of a tone carrying these cues from a background talker. These results confirm the importance of the F0 of target speech for EAS (in simulation). They indicate that significant benefit can be provided by a tone carrying F0 and amplitude envelope cues. The results support a glimpsing account of EAS and argue against segregation. PMID:19275323

  9. Acoustic and Articulatory Features of Diphthong Production: A Speech Clarity Study

    ERIC Educational Resources Information Center

    Tasko, Stephen M.; Greilick, Kristin

    2010-01-01

    Purpose: The purpose of this study was to evaluate how speaking clearly influences selected acoustic and orofacial kinematic measures associated with diphthong production. Method: Forty-nine speakers, drawn from the University of Wisconsin X-Ray Microbeam Speech Production Database (J. R. Westbury, 1994), served as participants. Samples of clear…

  10. Effects of contextual cues on speech recognition in simulated electric-acoustic stimulation.

    PubMed

    Kong, Ying-Yee; Donaldson, Gail; Somarowthu, Ala

    2015-05-01

    Low-frequency acoustic cues have shown to improve speech perception in cochlear-implant listeners. However, the mechanisms underlying this benefit are still not well understood. This study investigated the extent to which low-frequency cues can facilitate listeners' use of linguistic knowledge in simulated electric-acoustic stimulation (EAS). Experiment 1 examined differences in the magnitude of EAS benefit at the phoneme, word, and sentence levels. Speech materials were processed via noise-channel vocoding and lowpass (LP) filtering. The amount of spectral degradation in the vocoder speech was varied by applying different numbers of vocoder channels. Normal-hearing listeners were tested on vocoder-alone, LP-alone, and vocoder + LP conditions. Experiment 2 further examined factors that underlie the context effect on EAS benefit at the sentence level by limiting the low-frequency cues to temporal envelope and periodicity (AM + FM). Results showed that EAS benefit was greater for higher-context than for lower-context speech materials even when the LP ear received only low-frequency AM + FM cues. Possible explanations for the greater EAS benefit observed with higher-context materials may lie in the interplay between perceptual and expectation-driven processes for EAS speech recognition, and/or the band-importance functions for different types of speech materials. PMID:25994712

  11. Effects of contextual cues on speech recognition in simulated electric-acoustic stimulation

    PubMed Central

    Kong, Ying-Yee; Donaldson, Gail; Somarowthu, Ala

    2015-01-01

    Low-frequency acoustic cues have shown to improve speech perception in cochlear-implant listeners. However, the mechanisms underlying this benefit are still not well understood. This study investigated the extent to which low-frequency cues can facilitate listeners' use of linguistic knowledge in simulated electric-acoustic stimulation (EAS). Experiment 1 examined differences in the magnitude of EAS benefit at the phoneme, word, and sentence levels. Speech materials were processed via noise-channel vocoding and lowpass (LP) filtering. The amount of spectral degradation in the vocoder speech was varied by applying different numbers of vocoder channels. Normal-hearing listeners were tested on vocoder-alone, LP-alone, and vocoder + LP conditions. Experiment 2 further examined factors that underlie the context effect on EAS benefit at the sentence level by limiting the low-frequency cues to temporal envelope and periodicity (AM + FM). Results showed that EAS benefit was greater for higher-context than for lower-context speech materials even when the LP ear received only low-frequency AM + FM cues. Possible explanations for the greater EAS benefit observed with higher-context materials may lie in the interplay between perceptual and expectation-driven processes for EAS speech recognition, and/or the band-importance functions for different types of speech materials. PMID:25994712

  12. Contributions of electric and acoustic hearing to bimodal speech and music perception.

    PubMed

    Crew, Joseph D; Galvin, John J; Landsberger, David M; Fu, Qian-Jie

    2015-01-01

    Cochlear implant (CI) users have difficulty understanding speech in noisy listening conditions and perceiving music. Aided residual acoustic hearing in the contralateral ear can mitigate these limitations. The present study examined contributions of electric and acoustic hearing to speech understanding in noise and melodic pitch perception. Data was collected with the CI only, the hearing aid (HA) only, and both devices together (CI+HA). Speech reception thresholds (SRTs) were adaptively measured for simple sentences in speech babble. Melodic contour identification (MCI) was measured with and without a masker instrument; the fundamental frequency of the masker was varied to be overlapping or non-overlapping with the target contour. Results showed that the CI contributes primarily to bimodal speech perception and that the HA contributes primarily to bimodal melodic pitch perception. In general, CI+HA performance was slightly improved relative to the better ear alone (CI-only) for SRTs but not for MCI, with some subjects experiencing a decrease in bimodal MCI performance relative to the better ear alone (HA-only). Individual performance was highly variable, and the contribution of either device to bimodal perception was both subject- and task-dependent. The results suggest that individualized mapping of CIs and HAs may further improve bimodal speech and music perception. PMID:25790349

  13. Primary Progressive Apraxia of Speech: Clinical Features and Acoustic and Neurologic Correlates

    PubMed Central

    Strand, Edythe A.; Clark, Heather; Machulda, Mary; Whitwell, Jennifer L.; Josephs, Keith A.

    2015-01-01

    Purpose This study summarizes 2 illustrative cases of a neurodegenerative speech disorder, primary progressive apraxia of speech (AOS), as a vehicle for providing an overview of the disorder and an approach to describing and quantifying its perceptual features and some of its temporal acoustic attributes. Method Two individuals with primary progressive AOS underwent speech-language and neurologic evaluations on 2 occasions, ranging from 2.0 to 7.5 years postonset. Performance on several tests, tasks, and rating scales, as well as several acoustic measures, were compared over time within and between cases. Acoustic measures were compared with performance of control speakers. Results Both patients initially presented with AOS as the only or predominant sign of disease and without aphasia or dysarthria. The presenting features and temporal progression were captured in an AOS Rating Scale, an Articulation Error Score, and temporal acoustic measures of utterance duration, syllable rates per second, rates of speechlike alternating motion and sequential motion, and a pairwise variability index measure. Conclusions AOS can be the predominant manifestation of neurodegenerative disease. Clinical ratings of its attributes and acoustic measures of some of its temporal characteristics can support its diagnosis and help quantify its salient characteristics and progression over time. PMID:25654422

  14. Effective Communication for Academic Chairs. SUNY Series in Speech Communication.

    ERIC Educational Resources Information Center

    Hickson, Mark, III, Ed.; Stacks, Don W.

    This book presents 11 contributed papers which examine communication aspects of the department chair position in academia. It is noted that most academic department chairs are not trained in management skills, including communication strategies. After an introductory chapter by Christopher H. Spicer and Ann Q. Staton, the following papers are…

  15. The advantages of sound localization and speech perception of bilateral electric acoustic stimulation

    PubMed Central

    Moteki, Hideaki; Kitoh, Ryosuke; Tsukada, Keita; Iwasaki, Satoshi; Nishio, Shin-Ya

    2015-01-01

    Conclusion: Bilateral electric acoustic stimulation (EAS) effectively improved speech perception in noise and sound localization in patients with high-frequency hearing loss. Objective: To evaluate bilateral EAS efficacy of sound localization detection and speech perception in noise in two cases of high-frequency hearing loss. Methods: Two female patients, aged 38 and 45 years, respectively, received bilateral EAS sequentially. Pure-tone audiometry was performed preoperatively and postoperatively to evaluate the hearing preservation in the lower frequencies. Speech perception outcomes in quiet and noise and sound localization were assessed with unilateral and bilateral EAS. Results: Residual hearing in the lower frequencies was well preserved after insertion of a FLEX24 electrode (24 mm) using the round window approach. After bilateral EAS, speech perception improved in quiet and even more so in noise. In addition, the sound localization ability of both cases with bilateral EAS improved remarkably. PMID:25423260

  16. Vowel Acoustics in Dysarthria: Speech Disorder Diagnosis and Classification

    ERIC Educational Resources Information Center

    Lansford, Kaitlin L.; Liss, Julie M.

    2014-01-01

    Purpose: The purpose of this study was to determine the extent to which vowel metrics are capable of distinguishing healthy from dysarthric speech and among different forms of dysarthria. Method: A variety of vowel metrics were derived from spectral and temporal measurements of vowel tokens embedded in phrases produced by 45 speakers with…

  17. Contribution of Consonant Landmarks to Speech Recognition in Simulated Acoustic-Electric Hearing

    PubMed Central

    Chen, Fei; Loizou, Philipos C.

    2009-01-01

    Objectives The purpose of this study is to assess the contribution of information provided by obstruent consonants (e.g. stops, fricatives) to speech intelligibility in simulated acoustic-electric hearing. As a secondary objective, this study examines the performance of an objective measure that can potentially be used for predicting the intelligibility of vocoded speech. Design Noise-corrupted sentences are used in Exp. 1 in which the noise-corrupted obstruent consonants are replaced with clean obstruent consonants, while leaving the sonorant sounds (vowels, semivowels and nasals) corrupted. In one condition, listeners have only access to the low-frequency (<600 Hz) acoustic portion of the clean consonant spectra, in another condition listeners have only access to the higher frequency (>600 Hz) portion (vocoded) of the clean consonant spectra, and in the third condition they have access to both. In Exp. 2, we investigate a speech-coding strategy that selectively attenuates the low-frequency portion of the consonant spectra while leaving the vocoded portion corrupted by noise. Finally, using the data collected from Exp. 1 and 2, we evaluate the performance of an objective measure in terms of predicting intelligibility of vocoded speech. This measure was originally designed to predict speech quality and has never been evaluated with vocoded speech. Results Significant improvements (about 30 percentage points) in intelligibility were noted in Exp. 1 in steady and two-talker masker conditions when the listeners had access to the clean obstruent consonants in both the acoustic and vocoded portions of the spectrum. The improvement was more evident in the low SNR levels (−5 and 0 dB). Further analysis indicated that it was access to the vocoded portion of the consonant spectra, rather than access to the low-frequency acoustic portion of the consonant spectra that contributed the most to the large improvements in performance. In Exp. 2, a small (14 percentage points

  18. Fundamental frequency is critical to speech perception in noise in combined acoustic and electric hearinga

    PubMed Central

    Carroll, Jeff; Tiaden, Stephanie; Zeng, Fan-Gang

    2011-01-01

    Cochlear implant (CI) users have been shown to benefit from residual low-frequency hearing, specifically in pitch related tasks. It remains unclear whether this benefit is dependent on fundamental frequency (F0) or other acoustic cues. Three experiments were conducted to determine the role of F0, as well as its frequency modulated (FM) and amplitude modulated (AM) components, in speech recognition with a competing voice. In simulated CI listeners, the signal-to-noise ratio was varied to estimate the 50% correct response. Simulation results showed that the F0 cue contributes to a significant proportion of the benefit seen with combined acoustic and electric hearing, and additionally that this benefit is due to the FM rather than the AM component. In actual CI users, sentence recognition scores were collected with either the full F0 cue containing both the FM and AM components or the 500-Hz low-pass speech cue containing the F0 and additional harmonics. The F0 cue provided a benefit similar to the low-pass cue for speech in noise, but not in quiet. Poorer CI users benefited more from the F0 cue than better users. These findings suggest that F0 is critical to improving speech perception in noise in combined acoustic and electric hearing. PMID:21973360

  19. Multi-channel spatial auditory display for speech communications

    NASA Astrophysics Data System (ADS)

    Begault, Durand; Erbe, Tom

    1993-10-01

    A spatial auditory display for multiple speech communications was developed at NASA-Ames Research Center. Input is spatialized by use of simplified head-related transfer functions, adapted for FIR filtering on Motorola 56001 digital signal processors. Hardware and firmware design implementations are overviewed for the initial prototype developed for NASA-Kennedy Space Center. An adaptive staircase method was used to determine intelligibility levels of four letter call signs used by launch personnel at NASA, against diotic speech babble. Spatial positions at 30 deg azimuth increments were evaluated. The results from eight subjects showed a maximal intelligibility improvement of about 6 to 7 dB when the signal was spatialized to 60 deg or 90 deg azimuth positions.

  20. Multi-channel spatial auditory display for speech communications

    NASA Technical Reports Server (NTRS)

    Begault, Durand; Erbe, Tom

    1993-01-01

    A spatial auditory display for multiple speech communications was developed at NASA-Ames Research Center. Input is spatialized by use of simplified head-related transfer functions, adapted for FIR filtering on Motorola 56001 digital signal processors. Hardware and firmware design implementations are overviewed for the initial prototype developed for NASA-Kennedy Space Center. An adaptive staircase method was used to determine intelligibility levels of four letter call signs used by launch personnel at NASA, against diotic speech babble. Spatial positions at 30 deg azimuth increments were evaluated. The results from eight subjects showed a maximal intelligibility improvement of about 6 to 7 dB when the signal was spatialized to 60 deg or 90 deg azimuth positions.

  1. Effects of deafness on acoustic characteristics of American English tense/lax vowels in maternal speech to infants

    PubMed Central

    Kondaurova, Maria V.; Bergeson, Tonya R.; Dilley, Laura C.

    2012-01-01

    Recent studies have demonstrated that mothers exaggerate phonetic properties of infant-directed (ID) speech. However, these studies focused on a single acoustic dimension (frequency), whereas speech sounds are composed of multiple acoustic cues. Moreover, little is known about how mothers adjust phonetic properties of speech to children with hearing loss. This study examined mothers’ production of frequency and duration cues to the American English tense/lax vowel contrast in speech to profoundly deaf (N = 14) and normal-hearing (N = 14) infants, and to an adult experimenter. First and second formant frequencies and vowel duration of tense (/i/, /u/) and lax (/I/, /ʊ/) vowels were measured. Results demonstrated that for both infant groups mothers hyperarticulated the acoustic vowel space and increased vowel duration in ID speech relative to adult-directed speech. Mean F2 values were decreased for the /u/ vowel and increased for the /I/ vowel, and vowel duration was longer for the /i/, /u/, and /I/ vowels in ID speech. However, neither acoustic cue differed in speech to hearing-impaired or normal-hearing infants. These results suggest that both formant frequencies and vowel duration that differentiate American English tense/lx vowel contrasts are modified in ID speech regardless of the hearing status of the addressee. PMID:22894224

  2. A Chimpanzee Recognizes Synthetic Speech With Significantly Reduced Acoustic Cues to Phonetic Content

    PubMed Central

    Heimbauer, Lisa A.; Beran, Michael J.; Owren, Michael J.

    2011-01-01

    Summary A long-standing debate concerns whether humans are specialized for speech perception [1–7], which some researchers argue is demonstrated by the ability to understand synthetic speech with significantly reduced acoustic cues to phonetic content [2–4,7]. We tested a chimpanzee (Pan troglodytes) that recognizes 128 spoken words [8,9], asking whether she could understand such speech. Three experiments presented 48 individual words, with the animal selecting a corresponding visuo-graphic symbol from among four alternatives. Experiment 1 tested spectrally reduced, noise-vocoded (NV) synthesis, originally developed to simulate input received by human cochlear-implant users [10]. Experiment 2 tested “impossibly unspeechlike” [3] sine-wave (SW) synthesis, which reduces speech to just three moving tones [11]. Although receiving only intermittent and non-contingent reward, the chimpanzee performed well above chance level, including when hearing synthetic versions for the first time. Recognition of SW words was least accurate, but improved in Experiment 3 when natural words in the same session were rewarded. The chimpanzee was more accurate with NV than SW versions, as were 32 human participants hearing these items. The chimpanzee's ability to spontaneously recognize acoustically reduced synthetic words suggests that experience rather than specialization is critical for speech-perception capabilities that some have suggested are uniquely human [12–14]. PMID:21723125

  3. Formant Centralization Ratio (FCR): A proposal for a new acoustic measure of dysarthric speech

    PubMed Central

    Sapir, Shimon; Ramig, Lorraine O.; Spielman, Jennifer L.; Fox, Cynthia

    2009-01-01

    Background and Aims The vowel space area (VSA) has been used as an acoustic metric of dysarthric speech, but with varying degrees of success. Here we test an alternative metric -- Formant centralization ratio (FCR) -- that is hypothesized to more effectively differentiate dysarthric from healthy speech and register treatment effects. Methods Speech recordings of 38 individuals with idiopathic Parkinson disease (IPD) and dysarthria (19 of whom received one month of intensive speech therapy (LSVT® LOUD)) and 14 healthy controls were acoustically analyzed. Vowels were extracted from short phrases. The same vowel-formant elements were used to construct the FCR, expressed as (F2u+F2𝔞+F1i+F1u)/(F2i+F1𝔞), the VSA, expressed as ABS((F1i*(F2𝔞–F2u)+F1𝔞*(F2u–F2i)+F1u*(F2i–F2𝔞))/2), a logarithmically scaled version of the VSA (LnVSA), and the F2i/F2u ratio. Results Unlike the VSA and the LnVSA, the FCR and F2i/F2u robustly differentiated dysarthric from healthy speech and were not gender-sensitive. All metrics effectively registered treatment effects and were strongly correlated with each other. Conclusions Albeit preliminary, the present findings indicate that the FCR is a sensitive, valid and reliable acoustic metric for distinguishing dysarthric from normal speech and for monitoring treatment effects, probably so because of reduced sensitivity to inter-speaker variability and enhanced sensitivity to vowel centralization. PMID:19948755

  4. Understanding the Abstract Role of Speech in Communication at 12 Months

    ERIC Educational Resources Information Center

    Martin, Alia; Onishi, Kristine H.; Vouloumanos, Athena

    2012-01-01

    Adult humans recognize that even unfamiliar speech can communicate information between third parties, demonstrating an ability to separate communicative function from linguistic content. We examined whether 12-month-old infants understand that speech can communicate before they understand the meanings of specific words. Specifically, we test the…

  5. Increased pain intensity is associated with greater verbal communication difficulty and increased production of speech and co-speech gestures.

    PubMed

    Rowbotham, Samantha; Wardy, April J; Lloyd, Donna M; Wearden, Alison; Holler, Judith

    2014-01-01

    Effective pain communication is essential if adequate treatment and support are to be provided. Pain communication is often multimodal, with sufferers utilising speech, nonverbal behaviours (such as facial expressions), and co-speech gestures (bodily movements, primarily of the hands and arms that accompany speech and can convey semantic information) to communicate their experience. Research suggests that the production of nonverbal pain behaviours is positively associated with pain intensity, but it is not known whether this is also the case for speech and co-speech gestures. The present study explored whether increased pain intensity is associated with greater speech and gesture production during face-to-face communication about acute, experimental pain. Participants (N = 26) were exposed to experimentally elicited pressure pain to the fingernail bed at high and low intensities and took part in video-recorded semi-structured interviews. Despite rating more intense pain as more difficult to communicate (t(25)  = 2.21, p =  .037), participants produced significantly longer verbal pain descriptions and more co-speech gestures in the high intensity pain condition (Words: t(25)  = 3.57, p  = .001; Gestures: t(25)  = 3.66, p =  .001). This suggests that spoken and gestural communication about pain is enhanced when pain is more intense. Thus, in addition to conveying detailed semantic information about pain, speech and co-speech gestures may provide a cue to pain intensity, with implications for the treatment and support received by pain sufferers. Future work should consider whether these findings are applicable within the context of clinical interactions about pain. PMID:25343486

  6. Speech Recognition and Acoustic Features in Combined Electric and Acoustic Stimulation

    ERIC Educational Resources Information Center

    Yoon, Yang-soo; Li, Yongxin; Fu, Qian-Jie

    2012-01-01

    Purpose: In this study, the authors aimed to identify speech information processed by a hearing aid (HA) that is additive to information processed by a cochlear implant (CI) as a function of signal-to-noise ratio (SNR). Method: Speech recognition was measured with CI alone, HA alone, and CI + HA. Ten participants were separated into 2 groups; good…

  7. From prosodic structure to acoustic saliency: A fMRI investigation of speech rate, clarity, and emphasis

    NASA Astrophysics Data System (ADS)

    Golfinopoulos, Elisa

    Acoustic variability in fluent speech can arise at many stages in speech production planning and execution. For example, at the phonological encoding stage, the grouping of phonemes into syllables determines which segments are coarticulated and, by consequence, segment-level acoustic variation. Likewise phonetic encoding, which determines the spatiotemporal extent of articulatory gestures, will affect the acoustic detail of segments. Functional magnetic resonance imaging (fMRI) was used to measure brain activity of fluent adult speakers in four speaking conditions: fast, normal, clear, and emphatic (or stressed) speech. These speech manner changes typically result in acoustic variations that do not change the lexical or semantic identity of productions but do affect the acoustic saliency of phonemes, syllables and/or words. Acoustic responses recorded inside the scanner were assessed quantitatively using eight acoustic measures and sentence duration was used as a covariate of non-interest in the neuroimaging analysis. Compared to normal speech, emphatic speech was characterized acoustically by a greater difference between stressed and unstressed vowels in intensity, duration, and fundamental frequency, and neurally by increased activity in right middle premotor cortex and supplementary motor area, and bilateral primary sensorimotor cortex. These findings are consistent with right-lateralized motor planning of prosodic variation in emphatic speech. Clear speech involved an increase in average vowel and sentence durations and average vowel spacing, along with increased activity in left middle premotor cortex and bilateral primary sensorimotor cortex. These findings are consistent with an increased reliance on feedforward control, resulting in hyper-articulation, under clear as compared to normal speech. Fast speech was characterized acoustically by reduced sentence duration and average vowel spacing, and neurally by increased activity in left anterior frontal

  8. Acoustic MIMO communications in a very shallow water channel

    NASA Astrophysics Data System (ADS)

    Zhou, Yuehai; Cao, Xiuling; Tong, Feng

    2015-12-01

    Underwater acoustic channels pose significant difficulty for the development of high speed communication due to highly limited band-width as well as hostile multipath interference. Enlightened by rapid progress of multiple input multiple output (MIMO) technologies in wireless communication scenarios, MIMO systems offer a potential solution by enabling multiple spatially parallel communication channels to improve communication performance as well as capacity. For MIMO acoustic communications, deep sea channels offer substantial spatial diversity among multiple channels that can be exploited to address simultaneous multipath and co-channel interference. At the same time, there are increasing requirements for high speed underwater communication in very shallow water area (for example, a depth less than 10 m). In this paper, a space-time multichannel adaptive receiver consisting of multiple decision feedback equalizers (DFE) is adopted as the receiver for a very shallow water MIMO acoustic communication system. The performance of multichannel DFE receivers with relatively small number of receiving elements are analyzed and compared with that of the multichannel time reversal receiver to evaluate the impact of limited spatial diversity on multi-channel equalization and time reversal processing. The results of sea trials in a very shallow water channel are presented to demonstrate the feasibility of very shallow water MIMO acoustic communication.

  9. Coherent underwater acoustic communication: Channel modeling and receiver design

    NASA Astrophysics Data System (ADS)

    Song, Aijun; Badiey, Mohsen

    2012-11-01

    High frequency acoustic communication (8-50 kHz) has attracted much attention recently due to increasing scientific and commercial activities in the marine environment. Based on multiple at-sea experiments, we have developed a number of communication techniques to address the receiver design challenges, resulting in significant data rate increases in the dynamic ocean. Our studies also show that it is possible to simulate the acoustic communication channel for its intensity profile, temporal coherence, and Doppler spread, leading to realistic representations of the at-sea measurements. We present our recent results in these two aspects, receiver design and channel modeling, for the mentioned frequency band.

  10. Knowledge and attitudes of teachers regarding the impact of classroom acoustics on speech perception and learning.

    PubMed

    Ramma, Lebogang

    2009-01-01

    This study investigated the knowledge and attitude of primary school teachers regarding the impact of poor classroom acoustics on learners' speech perception and learning in class. Classrooms with excessive background noise and reflective surfaces could be a barrier to learning, and it is important that teachers are aware of this. There is currently limited research data about teachers' knowledge regarding the topic of classroom acoustics. Seventy teachers from three Johannesburg primary schools participated in this study. A survey by way of structured self-administered questionnaire was the primary data collection method. The findings of this study showed that most of the participants in this study did not have adequate knowledge of classroom acoustics. Most of the participants were also unaware of the impact that classrooms with poor acoustic environments can have on speech perception and learning. These results are discussed in relation to the practical implication of empowering teachers to manage the acoustic environment of their classrooms, limitations of the study as well as implications for future research.

  11. An eighth-scale speech source for subjective assessments in acoustic models

    NASA Astrophysics Data System (ADS)

    Orlowski, R. J.

    1981-08-01

    The design of a source is described which is suitable for making speech recordings in eighth-scale acoustic models of auditoria. An attempt was made to match the directionality of the source with the directionality of the human voice using data reported in the literature. A narrow aperture was required for the design which was provided by mounting an inverted conical horn over the diaphragm of a high frequency loudspeaker. Resonance problems were encountered with the use of a horn and a description is given of the electronic techniques adopted to minimize the effect of these resonances. Subjective and objective assessments on the completed speech source have proved satisfactory. It has been used in a modelling exercise concerned with the acoustic design of a theatre with a thrust-type stage.

  12. Acoustic-Phonetic Differences between Infant- and Adult-Directed Speech: The Role of Stress and Utterance Position

    ERIC Educational Resources Information Center

    Wang, Yuanyuan; Seidl, Amanda; Cristia, Alejandrina

    2015-01-01

    Previous studies have shown that infant-directed speech (IDS) differs from adult-directed speech (ADS) on a variety of dimensions. The aim of the current study was to investigate whether acoustic differences between IDS and ADS in English are modulated by prosodic structure. We compared vowels across the two registers (IDS, ADS) in both stressed…

  13. Speech after Radial Forearm Free Flap Reconstruction of the Tongue: A Longitudinal Acoustic Study of Vowel and Diphthong Sounds

    ERIC Educational Resources Information Center

    Laaksonen, Juha-Pertti; Rieger, Jana; Happonen, Risto-Pekka; Harris, Jeffrey; Seikaly, Hadi

    2010-01-01

    The purpose of this study was to use acoustic analyses to describe speech outcomes over the course of 1 year after radial forearm free flap (RFFF) reconstruction of the tongue. Eighteen Canadian English-speaking females and males with reconstruction for oral cancer had speech samples recorded (pre-operative, and 1 month, 6 months, and 1 year…

  14. Are you a good mimic? Neuro-acoustic signatures for speech imitation ability.

    PubMed

    Reiterer, Susanne M; Hu, Xiaochen; Sumathi, T A; Singh, Nandini C

    2013-01-01

    We investigated individual differences in speech imitation ability in late bilinguals using a neuro-acoustic approach. One hundred and thirty-eight German-English bilinguals matched on various behavioral measures were tested for "speech imitation ability" in a foreign language, Hindi, and categorized into "high" and "low ability" groups. Brain activations and speech recordings were obtained from 26 participants from the two extreme groups as they performed a functional neuroimaging experiment which required them to "imitate" sentences in three conditions: (A) German, (B) English, and (C) German with fake English accent. We used recently developed novel acoustic analysis, namely the "articulation space" as a metric to compare speech imitation abilities of the two groups. Across all three conditions, direct comparisons between the two groups, revealed brain activations (FWE corrected, p < 0.05) that were more widespread with significantly higher peak activity in the left supramarginal gyrus and postcentral areas for the low ability group. The high ability group, on the other hand showed significantly larger articulation space in all three conditions. In addition, articulation space also correlated positively with imitation ability (Pearson's r = 0.7, p < 0.01). Our results suggest that an expanded articulation space for high ability individuals allows access to a larger repertoire of sounds, thereby providing skilled imitators greater flexibility in pronunciation and language learning. PMID:24155739

  15. Are you a good mimic? Neuro-acoustic signatures for speech imitation ability

    PubMed Central

    Reiterer, Susanne M.; Hu, Xiaochen; Sumathi, T. A.; Singh, Nandini C.

    2013-01-01

    We investigated individual differences in speech imitation ability in late bilinguals using a neuro-acoustic approach. One hundred and thirty-eight German-English bilinguals matched on various behavioral measures were tested for “speech imitation ability” in a foreign language, Hindi, and categorized into “high” and “low ability” groups. Brain activations and speech recordings were obtained from 26 participants from the two extreme groups as they performed a functional neuroimaging experiment which required them to “imitate” sentences in three conditions: (A) German, (B) English, and (C) German with fake English accent. We used recently developed novel acoustic analysis, namely the “articulation space” as a metric to compare speech imitation abilities of the two groups. Across all three conditions, direct comparisons between the two groups, revealed brain activations (FWE corrected, p < 0.05) that were more widespread with significantly higher peak activity in the left supramarginal gyrus and postcentral areas for the low ability group. The high ability group, on the other hand showed significantly larger articulation space in all three conditions. In addition, articulation space also correlated positively with imitation ability (Pearson's r = 0.7, p < 0.01). Our results suggest that an expanded articulation space for high ability individuals allows access to a larger repertoire of sounds, thereby providing skilled imitators greater flexibility in pronunciation and language learning. PMID:24155739

  16. Are you a good mimic? Neuro-acoustic signatures for speech imitation ability.

    PubMed

    Reiterer, Susanne M; Hu, Xiaochen; Sumathi, T A; Singh, Nandini C

    2013-01-01

    We investigated individual differences in speech imitation ability in late bilinguals using a neuro-acoustic approach. One hundred and thirty-eight German-English bilinguals matched on various behavioral measures were tested for "speech imitation ability" in a foreign language, Hindi, and categorized into "high" and "low ability" groups. Brain activations and speech recordings were obtained from 26 participants from the two extreme groups as they performed a functional neuroimaging experiment which required them to "imitate" sentences in three conditions: (A) German, (B) English, and (C) German with fake English accent. We used recently developed novel acoustic analysis, namely the "articulation space" as a metric to compare speech imitation abilities of the two groups. Across all three conditions, direct comparisons between the two groups, revealed brain activations (FWE corrected, p < 0.05) that were more widespread with significantly higher peak activity in the left supramarginal gyrus and postcentral areas for the low ability group. The high ability group, on the other hand showed significantly larger articulation space in all three conditions. In addition, articulation space also correlated positively with imitation ability (Pearson's r = 0.7, p < 0.01). Our results suggest that an expanded articulation space for high ability individuals allows access to a larger repertoire of sounds, thereby providing skilled imitators greater flexibility in pronunciation and language learning.

  17. Using others' words: conversational use of reported speech by individuals with aphasia and their communication partners.

    PubMed

    Hengst, Julie A; Frame, Simone R; Neuman-Stritzel, Tiffany; Gannaway, Rachel

    2005-02-01

    Reported speech, wherein one quotes or paraphrases the speech of another, has been studied extensively as a set of linguistic and discourse practices. Researchers agree that reported speech is pervasive, found across languages, and used in diverse contexts. However, to date, there have been no studies of the use of reported speech among individuals with aphasia. Grounded in an interactional sociolinguistic perspective, the study presented here documents and analyzes the use of reported speech by 7 adults with mild to moderately severe aphasia and their routine communication partners. Each of the 7 pairs was videotaped in 4 everyday activities at home or around the community, yielding over 27 hr of conversational interaction for analysis. A coding scheme was developed that identified 5 types of explicitly marked reported speech: direct, indirect, projected, indexed, and undecided. Analysis of the data documented reported speech as a common discourse practice used successfully by the individuals with aphasia and their communication partners. All participants produced reported speech at least once, and across all observations the target pairs produced 400 reported speech episodes (RSEs), 149 by individuals with aphasia and 251 by their communication partners. For all participants, direct and indirect forms were the most prevalent (70% of RSEs). Situated discourse analysis of specific episodes of reported speech used by 3 of the pairs provides detailed portraits of the diverse interactional, referential, social, and discourse functions of reported speech and explores ways that the pairs used reported speech to successfully frame talk despite their ongoing management of aphasia. PMID:15934449

  18. An Acoustic Communication Technique of Nanorobot Swarms for Nanomedicine Applications.

    PubMed

    Loscrí, Valeria; Vegni, Anna Maria

    2015-09-01

    In this contribution, we present a communication paradigm among nanodevices, based on acoustic vibrations for medical applications. We consider a swarm of nanorobots able to communicate in a distributed and decentralized fashion, propelled in a biological environment (i.e., the human brain). Each nanorobot is intended to i) recognize a cancer cell, ii) destroy it, and then iii) forward information about the presence of cancer formation to other nanorobots, through acoustic signals. The choice of acoustic waves as communication mean is related to the application context, where it is not advisable either to use indiscriminate chemical substances or electromagnetic waves. The effectiveness of the proposed approach is assessed in terms of achievement of the objective (i.e., to destroy the majority of tumor cells), and the velocity of detection and destruction of cancer cells, through a comparison with other related techniques. PMID:25898028

  19. An Acoustic Communication Technique of Nanorobot Swarms for Nanomedicine Applications.

    PubMed

    Loscrí, Valeria; Vegni, Anna Maria

    2015-09-01

    In this contribution, we present a communication paradigm among nanodevices, based on acoustic vibrations for medical applications. We consider a swarm of nanorobots able to communicate in a distributed and decentralized fashion, propelled in a biological environment (i.e., the human brain). Each nanorobot is intended to i) recognize a cancer cell, ii) destroy it, and then iii) forward information about the presence of cancer formation to other nanorobots, through acoustic signals. The choice of acoustic waves as communication mean is related to the application context, where it is not advisable either to use indiscriminate chemical substances or electromagnetic waves. The effectiveness of the proposed approach is assessed in terms of achievement of the objective (i.e., to destroy the majority of tumor cells), and the velocity of detection and destruction of cancer cells, through a comparison with other related techniques.

  20. Acoustic-Emergent Phonology in the Amplitude Envelope of Child-Directed Speech.

    PubMed

    Leong, Victoria; Goswami, Usha

    2015-01-01

    When acquiring language, young children may use acoustic spectro-temporal patterns in speech to derive phonological units in spoken language (e.g., prosodic stress patterns, syllables, phonemes). Children appear to learn acoustic-phonological mappings rapidly, without direct instruction, yet the underlying developmental mechanisms remain unclear. Across different languages, a relationship between amplitude envelope sensitivity and phonological development has been found, suggesting that children may make use of amplitude modulation (AM) patterns within the envelope to develop a phonological system. Here we present the Spectral Amplitude Modulation Phase Hierarchy (S-AMPH) model, a set of algorithms for deriving the dominant AM patterns in child-directed speech (CDS). Using Principal Components Analysis, we show that rhythmic CDS contains an AM hierarchy comprising 3 core modulation timescales. These timescales correspond to key phonological units: prosodic stress (Stress AM, ~2 Hz), syllables (Syllable AM, ~5 Hz) and onset-rime units (Phoneme AM, ~20 Hz). We argue that these AM patterns could in principle be used by naïve listeners to compute acoustic-phonological mappings without lexical knowledge. We then demonstrate that the modulation statistics within this AM hierarchy indeed parse the speech signal into a primitive hierarchically-organised phonological system comprising stress feet (proto-words), syllables and onset-rime units. We apply the S-AMPH model to two other CDS corpora, one spontaneous and one deliberately-timed. The model accurately identified 72-82% (freely-read CDS) and 90-98% (rhythmically-regular CDS) stress patterns, syllables and onset-rime units. This in-principle demonstration that primitive phonology can be extracted from speech AMs is termed Acoustic-Emergent Phonology (AEP) theory. AEP theory provides a set of methods for examining how early phonological development is shaped by the temporal modulation structure of speech across

  1. Acoustic-Emergent Phonology in the Amplitude Envelope of Child-Directed Speech

    PubMed Central

    Leong, Victoria; Goswami, Usha

    2015-01-01

    When acquiring language, young children may use acoustic spectro-temporal patterns in speech to derive phonological units in spoken language (e.g., prosodic stress patterns, syllables, phonemes). Children appear to learn acoustic-phonological mappings rapidly, without direct instruction, yet the underlying developmental mechanisms remain unclear. Across different languages, a relationship between amplitude envelope sensitivity and phonological development has been found, suggesting that children may make use of amplitude modulation (AM) patterns within the envelope to develop a phonological system. Here we present the Spectral Amplitude Modulation Phase Hierarchy (S-AMPH) model, a set of algorithms for deriving the dominant AM patterns in child-directed speech (CDS). Using Principal Components Analysis, we show that rhythmic CDS contains an AM hierarchy comprising 3 core modulation timescales. These timescales correspond to key phonological units: prosodic stress (Stress AM, ~2 Hz), syllables (Syllable AM, ~5 Hz) and onset-rime units (Phoneme AM, ~20 Hz). We argue that these AM patterns could in principle be used by naïve listeners to compute acoustic-phonological mappings without lexical knowledge. We then demonstrate that the modulation statistics within this AM hierarchy indeed parse the speech signal into a primitive hierarchically-organised phonological system comprising stress feet (proto-words), syllables and onset-rime units. We apply the S-AMPH model to two other CDS corpora, one spontaneous and one deliberately-timed. The model accurately identified 72–82% (freely-read CDS) and 90–98% (rhythmically-regular CDS) stress patterns, syllables and onset-rime units. This in-principle demonstration that primitive phonology can be extracted from speech AMs is termed Acoustic-Emergent Phonology (AEP) theory. AEP theory provides a set of methods for examining how early phonological development is shaped by the temporal modulation structure of speech across

  2. Acoustic-Emergent Phonology in the Amplitude Envelope of Child-Directed Speech.

    PubMed

    Leong, Victoria; Goswami, Usha

    2015-01-01

    When acquiring language, young children may use acoustic spectro-temporal patterns in speech to derive phonological units in spoken language (e.g., prosodic stress patterns, syllables, phonemes). Children appear to learn acoustic-phonological mappings rapidly, without direct instruction, yet the underlying developmental mechanisms remain unclear. Across different languages, a relationship between amplitude envelope sensitivity and phonological development has been found, suggesting that children may make use of amplitude modulation (AM) patterns within the envelope to develop a phonological system. Here we present the Spectral Amplitude Modulation Phase Hierarchy (S-AMPH) model, a set of algorithms for deriving the dominant AM patterns in child-directed speech (CDS). Using Principal Components Analysis, we show that rhythmic CDS contains an AM hierarchy comprising 3 core modulation timescales. These timescales correspond to key phonological units: prosodic stress (Stress AM, ~2 Hz), syllables (Syllable AM, ~5 Hz) and onset-rime units (Phoneme AM, ~20 Hz). We argue that these AM patterns could in principle be used by naïve listeners to compute acoustic-phonological mappings without lexical knowledge. We then demonstrate that the modulation statistics within this AM hierarchy indeed parse the speech signal into a primitive hierarchically-organised phonological system comprising stress feet (proto-words), syllables and onset-rime units. We apply the S-AMPH model to two other CDS corpora, one spontaneous and one deliberately-timed. The model accurately identified 72-82% (freely-read CDS) and 90-98% (rhythmically-regular CDS) stress patterns, syllables and onset-rime units. This in-principle demonstration that primitive phonology can be extracted from speech AMs is termed Acoustic-Emergent Phonology (AEP) theory. AEP theory provides a set of methods for examining how early phonological development is shaped by the temporal modulation structure of speech across

  3. An Essential Curriculum for the Small, Comprehensive College Speech Communication Department.

    ERIC Educational Resources Information Center

    Griffin, Keith H.

    Observing that speech communication, by whatever term it is called, is now a popular major among the traditional liberal arts, this paper proposes an essential speech communication curriculum for the small, comprehensive college. The first part of the paper presents a rationale for the curriculum by exploring how the curriculum reflects (1) its…

  4. The Microcomputer as an Efficient and Versatile Speech-Output Communication Aid.

    ERIC Educational Resources Information Center

    Rushakoff, G. Evan

    The use of microcomputers as communication aids for physically handicapped children and adults is described. Characteristics of stationary speech output communication aids addressed include encoding by single switch users, an easy to change vocabulary through the use of an "editor," capability to speak via synthesized speech in addition to…

  5. A secure communication suite for underwater acoustic sensor networks.

    PubMed

    Dini, Gianluca; Lo Duca, Angelica

    2012-01-01

    In this paper we describe a security suite for Underwater Acoustic Sensor Networks comprising both fixed and mobile nodes. The security suite is composed of a secure routing protocol and a set of cryptographic primitives aimed at protecting the confidentiality and the integrity of underwater communication while taking into account the unique characteristics and constraints of the acoustic channel. By means of experiments and simulations based on real data, we show that the suite is suitable for an underwater networking environment as it introduces limited, and sometimes negligible, communication and power consumption overhead.

  6. A Secure Communication Suite for Underwater Acoustic Sensor Networks

    PubMed Central

    Dini, Gianluca; Duca, Angelica Lo

    2012-01-01

    In this paper we describe a security suite for Underwater Acoustic Sensor Networks comprising both fixed and mobile nodes. The security suite is composed of a secure routing protocol and a set of cryptographic primitives aimed at protecting the confidentiality and the integrity of underwater communication while taking into account the unique characteristics and constraints of the acoustic channel. By means of experiments and simulations based on real data, we show that the suite is suitable for an underwater networking environment as it introduces limited, and sometimes negligible, communication and power consumption overhead. PMID:23202204

  7. The Effects of Noise on Speech Recognition in Cochlear Implant Subjects: Predictions and Analysis Using Acoustic Models

    NASA Astrophysics Data System (ADS)

    Remus, Jeremiah J.; Collins, Leslie M.

    2005-12-01

    Cochlear implants can provide partial restoration of hearing, even with limited spectral resolution and loss of fine temporal structure, to severely deafened individuals. Studies have indicated that background noise has significant deleterious effects on the speech recognition performance of cochlear implant patients. This study investigates the effects of noise on speech recognition using acoustic models of two cochlear implant speech processors and several predictive signal-processing-based analyses. The results of a listening test for vowel and consonant recognition in noise are presented and analyzed using the rate of phonemic feature transmission for each acoustic model. Three methods for predicting patterns of consonant and vowel confusion that are based on signal processing techniques calculating a quantitative difference between speech tokens are developed and tested using the listening test results. Results of the listening test and confusion predictions are discussed in terms of comparisons between acoustic models and confusion prediction performance.

  8. RAPID ACOUSTIC PROCESSING IN THE AUDITORY BRAINSTEM IS NOT RELATED TO CORTICAL ASYMMETRY FOR THE SYLLABLE RATE OF SPEECH

    PubMed Central

    Abrams, Daniel A.; Nicol, Trent; Zecker, Steven; Kraus, Nina

    2010-01-01

    Objective Temporal acuity in the auditory brainstem is correlated with left-dominant patterns of cortical asymmetry for processing rapid speech-sound stimuli. Here we investigate whether a similar relationship exists between brainstem processing of rapid speech components and cortical processing of syllable patterns in speech. Methods We measured brainstem and cortical evoked potentials in response to speech tokens in 23 children. We used established measures of auditory brainstem and cortical activity to examine functional relationships between these structures. Results We found no relationship between brainstem responses to fast acoustic elements of speech and right-dominant cortical processing of syllable patterns. Conclusions Brainstem processing of rapid elements in speech is not functionally related to rightward cortical asymmetry associated with the processing of syllable-rate features in speech. Viewed together with previous evidence linking brainstem timing with leftward cortical asymmetry for faster acoustic features, findings support the existence of distinct mechanisms for encoding rapid vs. slow elements of speech. Significance Results provide a fundamental advance in our knowledge of the segregation of subcortical input associated with cortical asymmetries for acoustic rate processing in the human auditory system. Implications of these findings for auditory perception, reading ability and development are discussed. PMID:20378402

  9. Impact of Aberrant Acoustic Properties on the Perception of Sound Quality in Electrolarynx Speech

    ERIC Educational Resources Information Center

    Meltzner, Geoffrey S.; Hillman, Robert E.

    2005-01-01

    A large percentage of patients who have undergone laryngectomy to treat advanced laryngeal cancer rely on an electrolarynx (EL) to communicate verbally. Although serviceable, EL speech is plagued by shortcomings in both sound quality and intelligibility. This study sought to better quantify the relative contributions of previously identified…

  10. Comparative efficacy of the picture exchange communication system (PECS) versus a speech-generating device: effects on social-communicative skills and speech development.

    PubMed

    Boesch, Miriam C; Wendt, Oliver; Subramanian, Anu; Hsu, Ning

    2013-09-01

    The Picture Exchange Communication System (PECS) and a speech-generating device (SGD) were compared in a study with a multiple baseline, alternating treatment design. The effectiveness of these methods in increasing social-communicative behavior and natural speech production were assessed with three elementary school-aged children with severe autism who demonstrated extremely limited functional communication skills. Results for social-communicative behavior were mixed for all participants in both treatment conditions. Relatively little difference was observed between PECS and SGD conditions. Although findings were inconclusive, data patterns suggest that Phase II of the PECS training protocol is conducive to encouraging social-communicative behavior. Data for speech outcomes did not reveal any increases across participants, and no differences between treatment conditions were observed.

  11. Acoustic landmarks drive delta-theta oscillations to enable speech comprehension by facilitating perceptual parsing.

    PubMed

    Doelling, Keith B; Arnal, Luc H; Ghitza, Oded; Poeppel, David

    2014-01-15

    A growing body of research suggests that intrinsic neuronal slow (<10 Hz) oscillations in auditory cortex appear to track incoming speech and other spectro-temporally complex auditory signals. Within this framework, several recent studies have identified critical-band temporal envelopes as the specific acoustic feature being reflected by the phase of these oscillations. However, how this alignment between speech acoustics and neural oscillations might underpin intelligibility is unclear. Here we test the hypothesis that the 'sharpness' of temporal fluctuations in the critical band envelope acts as a temporal cue to speech syllabic rate, driving delta-theta rhythms to track the stimulus and facilitate intelligibility. We interpret our findings as evidence that sharp events in the stimulus cause cortical rhythms to re-align and parse the stimulus into syllable-sized chunks for further decoding. Using magnetoencephalographic recordings, we show that by removing temporal fluctuations that occur at the syllabic rate, envelope-tracking activity is reduced. By artificially reinstating these temporal fluctuations, envelope-tracking activity is regained. These changes in tracking correlate with intelligibility of the stimulus. Together, the results suggest that the sharpness of fluctuations in the stimulus, as reflected in the cochlear output, drive oscillatory activity to track and entrain to the stimulus, at its syllabic rate. This process likely facilitates parsing of the stimulus into meaningful chunks appropriate for subsequent decoding, enhancing perception and intelligibility.

  12. Acoustic landmarks drive delta-theta oscillations to enable speech comprehension by facilitating perceptual parsing

    PubMed Central

    Doelling, Keith; Arnal, Luc; Ghitza, Oded; Poeppel, David

    2013-01-01

    A growing body of research suggests that intrinsic neuronal slow (< 10 Hz) oscillations in auditory cortex appear to track incoming speech and other spectro-temporally complex auditory signals. Within this framework, several recent studies have identified critical-band temporal envelopes as the specific acoustic feature being reflected by the phase of these oscillations. However, how this alignment between speech acoustics and neural oscillations might underpin intelligibility is unclear. Here we test the hypothesis that the ‘sharpness’ of temporal fluctuations in the critical band envelope acts as a temporal cue to speech syllabic rate, driving delta-theta rhythms to track the stimulus and facilitate intelligibility. We interpret our findings as evidence that sharp events in the stimulus cause cortical rhythms to re-align and parse the stimulus into syllable-sized chunks for further decoding. Using magnetoencephalographic recordings, we show that by removing temporal fluctuations that occur at the syllabic rate, envelope-tracking activity is reduced. By artificially reinstating these temporal fluctuations, envelope-tracking activity is regained. These changes in tracking correlate with intelligibility of the stimulus. Together, the results suggest that the sharpness of fluctuations in the stimulus, as reflected in the cochlear output, drive oscillatory activity to track and entrain to the stimulus, at its syllabic rate. This process likely facilitates parsing of the stimulus into meaningful chunks appropriate for subsequent decoding, enhancing perception and intelligibility. PMID:23791839

  13. Control of spoken vowel acoustics and the influence of phonetic context in human speech sensorimotor cortex.

    PubMed

    Bouchard, Kristofer E; Chang, Edward F

    2014-09-17

    Speech production requires the precise control of vocal tract movements to generate individual speech sounds (phonemes) which, in turn, are rapidly organized into complex sequences. Multiple productions of the same phoneme can exhibit substantial variability, some of which is inherent to control of the vocal tract and its biomechanics, and some of which reflects the contextual effects of surrounding phonemes ("coarticulation"). The role of the CNS in these aspects of speech motor control is not well understood. To address these issues, we recorded multielectrode cortical activity directly from human ventral sensory-motor cortex (vSMC) during the production of consonant-vowel syllables. We analyzed the relationship between the acoustic parameters of vowels (pitch and formants) and cortical activity on a single-trial level. We found that vSMC activity robustly predicted acoustic parameters across vowel categories (up to 80% of variance), as well as different renditions of the same vowel (up to 25% of variance). Furthermore, we observed significant contextual effects on vSMC representations of produced phonemes that suggest active control of coarticulation: vSMC representations for vowels were biased toward the representations of the preceding consonant, and conversely, representations for consonants were biased toward upcoming vowels. These results reveal that vSMC activity for phonemes are not invariant and provide insight into the cortical mechanisms of coarticulation.

  14. Control of Spoken Vowel Acoustics and the Influence of Phonetic Context in Human Speech Sensorimotor Cortex

    PubMed Central

    Bouchard, Kristofer E.

    2014-01-01

    Speech production requires the precise control of vocal tract movements to generate individual speech sounds (phonemes) which, in turn, are rapidly organized into complex sequences. Multiple productions of the same phoneme can exhibit substantial variability, some of which is inherent to control of the vocal tract and its biomechanics, and some of which reflects the contextual effects of surrounding phonemes (“coarticulation”). The role of the CNS in these aspects of speech motor control is not well understood. To address these issues, we recorded multielectrode cortical activity directly from human ventral sensory-motor cortex (vSMC) during the production of consonant-vowel syllables. We analyzed the relationship between the acoustic parameters of vowels (pitch and formants) and cortical activity on a single-trial level. We found that vSMC activity robustly predicted acoustic parameters across vowel categories (up to 80% of variance), as well as different renditions of the same vowel (up to 25% of variance). Furthermore, we observed significant contextual effects on vSMC representations of produced phonemes that suggest active control of coarticulation: vSMC representations for vowels were biased toward the representations of the preceding consonant, and conversely, representations for consonants were biased toward upcoming vowels. These results reveal that vSMC activity for phonemes are not invariant and provide insight into the cortical mechanisms of coarticulation. PMID:25232105

  15. How private is your consultation? Acoustic and audiological measures of speech privacy in the otolaryngology clinic.

    PubMed

    Clamp, Philip J; Grant, David G; Zapala, David A; Hawkins, David B

    2011-01-01

    The right to confidentiality is a central tenet of the doctor-patient relationship. In the United Kingdom this right to confidentiality is recognised in published GMC guidance. In USA the Healthcare Insurance Portability and Accountability Act of 1996 (HIPAA) strengthened the legal requirement to protect patient information in all forms and failure to do so now constitutes a federal offence. The aims of this study are to assess the acoustic privacy of an otolaryngology outpatient consultation room. Acoustic privacy was measured using the articulation index (AI) and Bamford-Kowal-Bench (BKB) speech discrimination tests. BKB speech tests were calibrated to normal conversational volume (50 dB SPL). Both AI and BKB were calculated in four positions around the ENT clinic: within the consultation room, outside the consulting room door, in the nearest waiting area chair and in the farthest waiting area chair. Tests were undertaken with the clinic room door closed and open to assess the effect on privacy. With the clinic room door closed, mean BKB scores in nearest and farthest waiting area chairs were 51 and 41% respectively. AI scores in the waiting area chairs were 0.03 and 0.02. With the clinic room door open, privacy was lost in both AI and BKB testing, with almost 100% of word discernable at normal talking levels. The results of this study highlight the poor level of speech privacy within a standard ENT outpatient department. AI is a poor predictor or privacy.

  16. 30. LAUNCH CONTROL CAPSULE. ACOUSTICAL ENCLOSURE. OPERATORS' CHAIR AND COMMUNICATIONS ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    30. LAUNCH CONTROL CAPSULE. ACOUSTICAL ENCLOSURE. OPERATORS' CHAIR AND COMMUNICATIONS CONSOLE IN FOREGROUND. ELECTRONIC EQUIPMENT RACK AT LEFT; LAUNCH CONTROL CONSOLE WITH CAPTAIN JAMES L. KING, JR. IN CENTER. LIEUTENANT KEVIN R. MCCLUNEY IN BACKGROUND. VIEW TO SOUTHEAST. - Minuteman III ICBM Launch Control Facility November-1, 1.5 miles North of New Raymer & State Highway 14, New Raymer, Weld County, CO

  17. 27. LAUNCH CONTROL CAPSULE. ACOUSTICAL ENCLOSURE. COMMUNICATIONS CONSOLE AT LEFT; ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    27. LAUNCH CONTROL CAPSULE. ACOUSTICAL ENCLOSURE. COMMUNICATIONS CONSOLE AT LEFT; LAUNCH CONTROL CONSOLE AT RIGHT. PADLOCKED PANEL AT TOP CENTER CONTAINS MISSILE LAUNCH KEYS. SHOCK ISOLATOR AT FAR LEFT. VIEW TO EAST. - Minuteman III ICBM Launch Control Facility November-1, 1.5 miles North of New Raymer & State Highway 14, New Raymer, Weld County, CO

  18. Teachers and Teaching: Speech Production Accommodations Due to Changes in the Acoustic Environment

    PubMed Central

    Hunter, Eric J.; Bottalico, Pasquale; Graetzer, Simone; Leishman, Timothy W.; Berardi, Mark L.; Eyring, Nathan G.; Jensen, Zachary R.; Rolins, Michael K.; Whiting, Jennifer K.

    2016-01-01

    School teachers have an elevated risk of voice problems due to the vocal demands in the workplace. This manuscript presents the results of three studies investigating teachers’ voice use at work. In the first study, 57 teachers were observed for 2 weeks (waking hours) to compare how they used their voice in the school environment and in non-school environments. In a second study, 45 participants performed a short vocal task in two different rooms: a variable acoustic room and an anechoic chamber. Subjects were taken back and forth between the two rooms. Each time they entered the variable acoustics room, the reverberation time and/or the background noise condition had been modified. In this latter study, subjects responded to questions about their vocal comfort and their perception of changes in the acoustic environment. In a third study, 20 untrained vocalists performed a simple vocal task in the following conditions: with and without background babble and with and without transparent plexiglass shields to increase the first reflection. Relationships were examined between [1] the results for the room acoustic parameters; [2] the subjects’ perception of the room; and [3] the recorded speech acoustic. Several differences between male and female subjects were found; some of those differences held for each room condition (at school vs. not at school, reverberation level, noise level, and early reflection). PMID:26949426

  19. Effect of several acoustic cues on perceiving Mandarin retroflex affricates and fricatives in continuous speech.

    PubMed

    Zhu, Jian; Chen, Yaping

    2016-07-01

    Relatively little attention has been paid to the perception of the three-way contrast between unaspirated affricates, aspirated affricates and fricatives in Mandarin Chinese. This study reports two experiments that explore the acoustic cues relevant to the contrast between the Mandarin retroflex series /tʂ/, /tʂ(h)/ and /ʂ/ in continuous speech. Twenty participants performed two three-alternative forced-choice tasks, in which acoustic cues including closure, frication duration (FD), aspiration, and vocalic contexts (VCs) were systematically manipulated and presented in a carrier phrase. A subsequent classification tree analysis shows that FD distinguishes /tʂ/ from /tʂ(h)/ and /ʂ/, and that closure cues the affricate manner. Interactions between VC and individual cues are also found. The FD threshold for separating /ʂ/ and /tʂ/ is susceptible to the influence of the following vocalic segments, shifting to lower values if frication is followed by the low vowel /a/. On the other hand, while aspiration cues /tʂ(h)/ before /a/ and //, this acoustic cue is obscured by gesture continuation when /tʂ(h)/ precedes its homorganic approximant /ɻ/ in natural speech, which might cause potential confusion between /tʂ(h)/ and /ʂ/. PMID:27475170

  20. Logopenic and nonfluent variants of primary progressive aphasia are differentiated by acoustic measures of speech production.

    PubMed

    Ballard, Kirrie J; Savage, Sharon; Leyton, Cristian E; Vogel, Adam P; Hornberger, Michael; Hodges, John R

    2014-01-01

    Differentiation of logopenic (lvPPA) and nonfluent/agrammatic (nfvPPA) variants of Primary Progressive Aphasia is important yet remains challenging since it hinges on expert based evaluation of speech and language production. In this study acoustic measures of speech in conjunction with voxel-based morphometry were used to determine the success of the measures as an adjunct to diagnosis and to explore the neural basis of apraxia of speech in nfvPPA. Forty-one patients (21 lvPPA, 20 nfvPPA) were recruited from a consecutive sample with suspected frontotemporal dementia. Patients were diagnosed using the current gold-standard of expert perceptual judgment, based on presence/absence of particular speech features during speaking tasks. Seventeen healthy age-matched adults served as controls. MRI scans were available for 11 control and 37 PPA cases; 23 of the PPA cases underwent amyloid ligand PET imaging. Measures, corresponding to perceptual features of apraxia of speech, were periods of silence during reading and relative vowel duration and intensity in polysyllable word repetition. Discriminant function analyses revealed that a measure of relative vowel duration differentiated nfvPPA cases from both control and lvPPA cases (r(2) = 0.47) with 88% agreement with expert judgment of presence of apraxia of speech in nfvPPA cases. VBM analysis showed that relative vowel duration covaried with grey matter intensity in areas critical for speech motor planning and programming: precentral gyrus, supplementary motor area and inferior frontal gyrus bilaterally, only affected in the nfvPPA group. This bilateral involvement of frontal speech networks in nfvPPA potentially affects access to compensatory mechanisms involving right hemisphere homologues. Measures of silences during reading also discriminated the PPA and control groups, but did not increase predictive accuracy. Findings suggest that a measure of relative vowel duration from of a polysyllable word repetition task

  1. Logopenic and Nonfluent Variants of Primary Progressive Aphasia Are Differentiated by Acoustic Measures of Speech Production

    PubMed Central

    Ballard, Kirrie J.; Savage, Sharon; Leyton, Cristian E.; Vogel, Adam P.; Hornberger, Michael; Hodges, John R.

    2014-01-01

    Differentiation of logopenic (lvPPA) and nonfluent/agrammatic (nfvPPA) variants of Primary Progressive Aphasia is important yet remains challenging since it hinges on expert based evaluation of speech and language production. In this study acoustic measures of speech in conjunction with voxel-based morphometry were used to determine the success of the measures as an adjunct to diagnosis and to explore the neural basis of apraxia of speech in nfvPPA. Forty-one patients (21 lvPPA, 20 nfvPPA) were recruited from a consecutive sample with suspected frontotemporal dementia. Patients were diagnosed using the current gold-standard of expert perceptual judgment, based on presence/absence of particular speech features during speaking tasks. Seventeen healthy age-matched adults served as controls. MRI scans were available for 11 control and 37 PPA cases; 23 of the PPA cases underwent amyloid ligand PET imaging. Measures, corresponding to perceptual features of apraxia of speech, were periods of silence during reading and relative vowel duration and intensity in polysyllable word repetition. Discriminant function analyses revealed that a measure of relative vowel duration differentiated nfvPPA cases from both control and lvPPA cases (r2 = 0.47) with 88% agreement with expert judgment of presence of apraxia of speech in nfvPPA cases. VBM analysis showed that relative vowel duration covaried with grey matter intensity in areas critical for speech motor planning and programming: precentral gyrus, supplementary motor area and inferior frontal gyrus bilaterally, only affected in the nfvPPA group. This bilateral involvement of frontal speech networks in nfvPPA potentially affects access to compensatory mechanisms involving right hemisphere homologues. Measures of silences during reading also discriminated the PPA and control groups, but did not increase predictive accuracy. Findings suggest that a measure of relative vowel duration from of a polysyllable word repetition task

  2. Multi-carrier Communications over Time-varying Acoustic Channels

    NASA Astrophysics Data System (ADS)

    Aval, Yashar M.

    Acoustic communication is an enabling technology for many autonomous undersea systems, such as those used for ocean monitoring, offshore oil and gas industry, aquaculture, or port security. There are three main challenges in achieving reliable high-rate underwater communication: the bandwidth of acoustic channels is extremely limited, the propagation delays are long, and the Doppler distortions are more pronounced than those found in wireless radio channels. In this dissertation we focus on assessing the fundamental limitations of acoustic communication, and designing efficient signal processing methods that cam overcome these limitations. We address the fundamental question of acoustic channel capacity (achievable rate) for single-input-multi-output (SIMO) acoustic channels using a per-path Rician fading model, and focusing on two scenarios: narrowband channels where the channel statistics can be approximated as frequency- independent, and wideband channels where the nominal path loss is frequency-dependent. In each scenario, we compare several candidate power allocation techniques, and show that assigning uniform power across all frequencies for the first scenario, and assigning uniform power across a selected frequency-band for the second scenario, are the best practical choices in most cases, because the long propagation delay renders the feedback information outdated for power allocation based on the estimated channel response. We quantify our results using the channel information extracted form the 2010 Mobile Acoustic Communications Experiment (MACE'10). Next, we focus on achieving reliable high-rate communication over underwater acoustic channels. Specifically, we investigate orthogonal frequency division multiplexing (OFDM) as the state-of-the-art technique for dealing with frequency-selective multipath channels, and propose a class of methods that compensate for the time-variation of the underwater acoustic channel. These methods are based on multiple

  3. The role of metrical information in apraxia of speech. Perceptual and acoustic analyses of word stress.

    PubMed

    Aichert, Ingrid; Späth, Mona; Ziegler, Wolfram

    2016-02-01

    Several factors are known to influence speech accuracy in patients with apraxia of speech (AOS), e.g., syllable structure or word length. However, the impact of word stress has largely been neglected so far. More generally, the role of prosodic information at the phonetic encoding stage of speech production often remains unconsidered in models of speech production. This study aimed to investigate the influence of word stress on error production in AOS. Two-syllabic words with stress on the first (trochees) vs. the second syllable (iambs) were compared in 14 patients with AOS, three of them exhibiting pure AOS, and in a control group of six normal speakers. The patients produced significantly more errors on iambic than on trochaic words. A most prominent metrical effect was obtained for segmental errors. Acoustic analyses of word durations revealed a disproportionate advantage of the trochaic meter in the patients relative to the healthy controls. The results indicate that German apraxic speakers are sensitive to metrical information. It is assumed that metrical patterns function as prosodic frames for articulation planning, and that the regular metrical pattern in German, the trochaic form, has a facilitating effect on word production in patients with AOS. PMID:26792367

  4. Intonation and communicative intent in mothers' speech to infants: is the melody the message?

    PubMed

    Fernald, A

    1989-12-01

    This study explores the power of intonation to convey meaningful information about the communicative intent of the speaker in speech addressed to preverbal infants and in speech addressed to adults. Natural samples of infant- and adult-directed speech were recorded from 5 mothers of 12-month-old infants, in 5 standardized interactional contexts: Attention-bid, Approval, Prohibition, Comfort, and Game/Telephone. 25 infant-directed and 25 adult-directed vocalizations were electronically filtered to eliminate linguistic content. The content-filtered speech stimuli were presented to 80 adult subjects: 40 experienced parents and 40 students inexperienced with infants. The subjects' task was to identify the communicative intent of the speaker using only prosodic information, given a 5-alternative forced choice. Listeners were able to use intonation to identify the speaker's intent with significantly higher accuracy in infant-directed speech than in adult-directed speech. These findings suggest that the prosodic patterns of speech to infants are more informative than those of adult-adult speech, and may provide the infant with reliable cues to the communicative intent of the speaker. The interpretation of these results proposed here is that the relation of prosodic form to communicative function is made uniquely salient in the melodies of mothers' speech, and that these characteristic prosodic patterns are potentially meaningful to the preverbal infant.

  5. Proceedings of the Speech Communication Association Summer Conference: Mini Courses in Speech Communication (7th, Chicago, July 8-10, 1971).

    ERIC Educational Resources Information Center

    Jeffrey, Robert C., Ed.

    The Speech Communication Association's 1971 summer conference provided instruction in the application of basic research and innovative practices in communication. It was designed to assist elementary, secondary, and college teachers in the enrichment of content and procedures. The proceedings include syllabi, course units, and bibliographic…

  6. Frequency diversity for OFDM mobile communication via underwater acoustic channels

    NASA Astrophysics Data System (ADS)

    Qiao, Gang; Wang, Wei; Guo, Ran; Khan, Rehan; Wang, Yue

    2012-03-01

    The major constraint on the performance of orthogonal frequency division multiplexing (OFDM) based underwater acoustic (UWA) communication is to keep subcarriers orthogonal. In this paper, Doppler estimation and the respective compensation technique along with various diversity techniques were deliberated for OFDM-based systems best suited for underwater wireless information exchange. In practice, for mobile communication, adjustment and tuning of transducers in order to get spatial diversity is extremely difficult. Considering the relatively low coherence bandwidth in UWA, the frequency diversity design with the Doppler compensation function was elaborated here. The outfield experiments of mobile underwater acoustic communication (UWAC) based on OFDM were carried out with 0.17 bit/(s·Hz) spectral efficiency. The validity and the dependability of the scheme were also analyzed.

  7. The effects of selective attention and speech acoustics on neural speech-tracking in a multi-talker scene.

    PubMed

    Rimmele, Johanna M; Zion Golumbic, Elana; Schröger, Erich; Poeppel, David

    2015-07-01

    Attending to one speaker in multi-speaker situations is challenging. One neural mechanism proposed to underlie the ability to attend to a particular speaker is phase-locking of low-frequency activity in auditory cortex to speech's temporal envelope ("speech-tracking"), which is more precise for attended speech. However, it is not known what brings about this attentional effect, and specifically if it reflects enhanced processing of the fine structure of attended speech. To investigate this question we compared attentional effects on speech-tracking of natural versus vocoded speech which preserves the temporal envelope but removes the fine structure of speech. Pairs of natural and vocoded speech stimuli were presented concurrently and participants attended to one stimulus and performed a detection task while ignoring the other stimulus. We recorded magnetoencephalography (MEG) and compared attentional effects on the speech-tracking response in auditory cortex. Speech-tracking of natural, but not vocoded, speech was enhanced by attention, whereas neural tracking of ignored speech was similar for natural and vocoded speech. These findings suggest that the more precise speech-tracking of attended natural speech is related to processing its fine structure, possibly reflecting the application of higher-order linguistic processes. In contrast, when speech is unattended its fine structure is not processed to the same degree and thus elicits less precise speech-tracking more similar to vocoded speech.

  8. Lexico-semantic and acoustic-phonetic processes in the perception of noise-vocoded speech: implications for cochlear implantation

    PubMed Central

    McGettigan, Carolyn; Rosen, Stuart; Scott, Sophie K.

    2014-01-01

    Noise-vocoding is a transformation which, when applied to speech, severely reduces spectral resolution and eliminates periodicity, yielding a stimulus that sounds “like a harsh whisper” (Scott et al., 2000, p. 2401). This process simulates a cochlear implant, where the activity of many thousand hair cells in the inner ear is replaced by direct stimulation of the auditory nerve by a small number of tonotopically-arranged electrodes. Although a cochlear implant offers a powerful means of restoring some degree of hearing to profoundly deaf individuals, the outcomes for spoken communication are highly variable (Moore and Shannon, 2009). Some variability may arise from differences in peripheral representation (e.g., the degree of residual nerve survival) but some may reflect differences in higher-order linguistic processing. In order to explore this possibility, we used noise-vocoding to explore speech recognition and perceptual learning in normal-hearing listeners tested across several levels of the linguistic hierarchy: segments (consonants and vowels), single words, and sentences. Listeners improved significantly on all tasks across two test sessions. In the first session, individual differences analyses revealed two independently varying sources of variability: one lexico-semantic in nature and implicating the recognition of words and sentences, and the other an acoustic-phonetic factor associated with words and segments. However, consequent to learning, by the second session there was a more uniform covariance pattern concerning all stimulus types. A further analysis of phonetic feature recognition allowed greater insight into learning-related changes in perception and showed that, surprisingly, participants did not make full use of cues that were preserved in the stimuli (e.g., vowel duration). We discuss these findings in relation cochlear implantation, and suggest auditory training strategies to maximize speech recognition performance in the absence of

  9. Overview of requirements and networks for voice communications and speech processing

    NASA Astrophysics Data System (ADS)

    Ince, A. Nejat

    1990-05-01

    The use of voice for military and civil communications are discussed. The military operational requirements are outlined in relation to air operations, including the effects of propagational factors and electronic warfare. Structures of the existing NATO communications network and the evolving Integrated Service Digital Network (ISDN) are reviewed to show how they meet the requirements. It is concluded that speech coding at low-bit rates is a growing need for transmitting speech messages with a high level of security and reliability over low data-rate channels and for memory-efficient systems for voice storage, voice response, and voice mail. Furthermore, it is pointed out that the low-bit rate voice coding can ease the transition to shared channels for voice and data and can readily adopt voice messages for packet switching. The speech processing techniques and systems are then outlined as an introduction to the lectures of this series in terms of: the character of the speech signal, its generation and perception; speech coding which is mainly concerned with man-to-man voice communication; speech synthesis which deals with machine-to-man communication; speech recognition which is related to man-to-machine communication; and quality assessment of speech system and standards.

  10. Time Reversal Acoustic Communication Using Filtered Multitone Modulation.

    PubMed

    Sun, Lin; Chen, Baowei; Li, Haisen; Zhou, Tian; Li, Ruo

    2015-01-01

    The multipath spread in underwater acoustic channels is severe and, therefore, when the symbol rate of the time reversal (TR) acoustic communication using single-carrier (SC) modulation is high, the large intersymbol interference (ISI) span caused by multipath reduces the performance of the TR process and needs to be removed using the long adaptive equalizer as the post-processor. In this paper, a TR acoustic communication method using filtered multitone (FMT) modulation is proposed in order to reduce the residual ISI in the processed signal using TR. In the proposed method, FMT modulation is exploited to modulate information symbols onto separate subcarriers with high spectral containment and TR technique, as well as adaptive equalization is adopted at the receiver to suppress ISI and noise. The performance of the proposed method is assessed through simulation and real data from a trial in an experimental pool. The proposed method was compared with the TR acoustic communication using SC modulation with the same spectral efficiency. Results demonstrate that the proposed method can improve the performance of the TR process and reduce the computational complexity of adaptive equalization for post-process.

  11. Time Reversal Acoustic Communication Using Filtered Multitone Modulation

    PubMed Central

    Sun, Lin; Chen, Baowei; Li, Haisen; Zhou, Tian; Li, Ruo

    2015-01-01

    The multipath spread in underwater acoustic channels is severe and, therefore, when the symbol rate of the time reversal (TR) acoustic communication using single-carrier (SC) modulation is high, the large intersymbol interference (ISI) span caused by multipath reduces the performance of the TR process and needs to be removed using the long adaptive equalizer as the post-processor. In this paper, a TR acoustic communication method using filtered multitone (FMT) modulation is proposed in order to reduce the residual ISI in the processed signal using TR. In the proposed method, FMT modulation is exploited to modulate information symbols onto separate subcarriers with high spectral containment and TR technique, as well as adaptive equalization is adopted at the receiver to suppress ISI and noise. The performance of the proposed method is assessed through simulation and real data from a trial in an experimental pool. The proposed method was compared with the TR acoustic communication using SC modulation with the same spectral efficiency. Results demonstrate that the proposed method can improve the performance of the TR process and reduce the computational complexity of adaptive equalization for post-process. PMID:26393586

  12. Advantages from bilateral hearing in speech perception in noise with simulated cochlear implants and residual acoustic hearing.

    PubMed

    Schoof, Tim; Green, Tim; Faulkner, Andrew; Rosen, Stuart

    2013-02-01

    Acoustic simulations were used to study the contributions of spatial hearing that may arise from combining a cochlear implant with either a second implant or contralateral residual low-frequency acoustic hearing. Speech reception thresholds (SRTs) were measured in twenty-talker babble. Spatial separation of speech and noise was simulated using a spherical head model. While low-frequency acoustic information contralateral to the implant simulation produced substantially better SRTs there was no effect of spatial cues on SRT, even when interaural differences were artificially enhanced. Simulated bilateral implants showed a significant head shadow effect, but no binaural unmasking based on interaural time differences, and weak, inconsistent overall spatial release from masking. There was also a small but significant non-spatial summation effect. It appears that typical cochlear implant speech processing strategies may substantially reduce the utility of spatial cues, even in the absence of degraded neural processing arising from auditory deprivation. PMID:23363118

  13. The effects of selective attention and speech acoustics on neural speech-tracking in a multi-talker scene

    PubMed Central

    Rimmele, Johanna M.; Golumbic, Elana Zion; Schröger, Erich; Poeppel, David

    2015-01-01

    Attending to one speaker in multi-speaker situations is challenging. One neural mechanism proposed to underlie the ability to attend to a particular speaker is phase-locking of low-frequency activity in auditory cortex to speech’s temporal envelope (“speech-tracking”), which is more precise for attended speech. However, it is not known what brings about this attentional effect, and specifically if it reflects enhanced processing of the fine structure of attended speech. To investigate this question we compared attentional effects on speech-tracking of natural vs. vocoded speech which preserves the temporal envelope but removes the fine-structure of speech. Pairs of natural and vocoded speech stimuli were presented concurrently and participants attended to one stimulus and performed a detection task while ignoring the other stimulus. We recorded magnetoencephalography (MEG) and compared attentional effects on the speech-tracking response in auditory cortex. Speech-tracking of natural, but not vocoded, speech was enhanced by attention, whereas neural tracking of ignored speech was similar for natural and vocoded speech. These findings suggest that the more precise speech tracking of attended natural speech is related to processing its fine structure, possibly reflecting the application of higher-order linguistic processes. In contrast, when speech is unattended its fine structure is not processed to the same degree and thus elicits less precise speech tracking more similar to vocoded speech. PMID:25650107

  14. Advances in speech processing

    NASA Astrophysics Data System (ADS)

    Ince, A. Nejat

    1992-10-01

    The field of speech processing is undergoing a rapid growth in terms of both performance and applications and this is fueled by the advances being made in the areas of microelectronics, computation, and algorithm design. The use of voice for civil and military communications is discussed considering advantages and disadvantages including the effects of environmental factors such as acoustic and electrical noise and interference and propagation. The structure of the existing NATO communications network and the evolving Integrated Services Digital Network (ISDN) concept are briefly reviewed to show how they meet the present and future requirements. The paper then deals with the fundamental subject of speech coding and compression. Recent advances in techniques and algorithms for speech coding now permit high quality voice reproduction at remarkably low bit rates. The subject of speech synthesis is next treated where the principle objective is to produce natural quality synthetic speech from unrestricted text input. Speech recognition where the ultimate objective is to produce a machine which would understand conversational speech with unrestricted vocabulary, from essentially any talker, is discussed. Algorithms for speech recognition can be characterized broadly as pattern recognition approaches and acoustic phonetic approaches. To date, the greatest degree of success in speech recognition has been obtained using pattern recognition paradigms. It is for this reason that the paper is concerned primarily with this technique.

  15. Enhancing the Speech and Language Development of Communicatively Disordered Children through Music and Movement.

    ERIC Educational Resources Information Center

    Leung, Katherine

    The paper examines the suprasegmental aspects of speech and focuses on the benefits of music and movement to facilitate language development in handicapped children. The paper discusses the current thinking of suprasegmental functions of intonation, stress, and rhythm as the foundation for the organization of speech communication. Strategies for…

  16. Effects of Noise and Filtering on the Intelligibility of Speech Produced during Simultaneous Communication

    ERIC Educational Resources Information Center

    MacKenzie, Douglas J.; Schiavetti, Nicholas; Whitehead, Robert L.; Metz, Dale Evan

    2004-01-01

    This study investigated the effects of noise and filtering on the intelligibility of speech produced during simultaneous communication (SC). Four normal hearing, experienced sign language users were recorded under SC and speech alone (SA) conditions speaking Boothroyd's forced-choice phonetic contrast material designed for measurement of speech…

  17. Selected Speeches on Obscenity by Federal Communications Commission Chairman Dean Burch, 1969-74.

    ERIC Educational Resources Information Center

    Hartenberger, Karen Schmidt

    This study is a descriptive/historical account focusing on the obscenity issue and the selected manuscript speeches of Dean Burch while he served as chairman of the Federal Communications Commission (FCC) from October 1969 to March 1974. Research centers on the speaker and the specific manuscript speeches, considering the timeliness and…

  18. Information Sources in Hearing, Speech, and Communication Disorders. Part I, Publications, Developmental Edition.

    ERIC Educational Resources Information Center

    Lunin, Lois F., Ed.

    This guide to recent sources of published information in hearing, speech, and communication disorders lists 759 references. The items are arranged in nine major sections and are annotated except when reasons of processing prevent. The section on reviews covers hearing, language, speech, and additional reviews; the section on special serials…

  19. Microcomputing in Speech Communication: A Report of the Task Force on Use of Computers.

    ERIC Educational Resources Information Center

    Cummings, H. Wayland, Ed.

    The essays in this volume explore the ways the Speech Communication Association (SCA) might use the computing technology of the microprocessor. The six chapters focus on the following topics: (1) ways the new technology will impact on the processing of manuscripts for speech journals and the SCA's national conventions, with particular emphasis on…

  20. Marshall McLuhan: His Significance for the Field of Speech Communication.

    ERIC Educational Resources Information Center

    Ehninger, Douglas

    1969-01-01

    Speech as a mode of communication and speech education as a means of training for proficiency in the use of this mode are assuming increased importance in light of McLuhan's assertion that electronic media is replacing print as the main source of information and entertainment. Emphasis upon the uniqueness and independence of the oral mode, a…

  1. Suprasegmental Characteristics of Speech Produced during Simultaneous Communication by Inexperienced Signers

    ERIC Educational Resources Information Center

    Whitehead, Robert L.; Metz, Dale E.; Girardi, Erin; Irwin, Jacqueline; Krigsman, Amanda; Swanson, Christina; Mackenzie, Douglas; Schiavetti, Nicholas

    2009-01-01

    This study investigated suprasegmental variables of syllable stress and intonation contours in contextual speech produced during simultaneous communication (SC) by inexperienced signers. Ten hearing inexperienced sign language users were recorded under SC and speech-alone (SA) conditions speaking a set of sentences containing stressed versus…

  2. Acoustic system for communication in pipelines

    SciTech Connect

    Martin, II, Louis Peter; Cooper, John F.

    2008-09-09

    A system for communication in a pipe, or pipeline, or network of pipes containing a fluid. The system includes an encoding and transmitting sub-system connected to the pipe, or pipeline, or network of pipes that transmits a signal in the frequency range of 3-100 kHz into the pipe, or pipeline, or network of pipes containing a fluid, and a receiver and processor sub-system connected to the pipe, or pipeline, or network of pipes containing a fluid that receives said signal and uses said signal for a desired application.

  3. Asynchronous multiuser underwater acoustic communications (L).

    PubMed

    Cho, S E; Song, H C; Hodgkiss, W S

    2012-07-01

    An asynchronous multiuser system is proposed to support multiple-access underwater communications without the use of code-division multiple-access or a feedback channel. The rich multipath channels experienced by spatially separated users will be sufficient to ensure separation of collided packets at the base station. The iterative receiver will employ a combination of adaptive time-reversal processing, matching pursuit, and successive interference cancellation in a block-wise fashion to achieve multiuser separability. Data collected during the KAM11 experiment are used to illustrate the system's capability in a dynamic, time-varying environment.

  4. Competency-Based Teacher Education; Group 1 Report of the Education Priorities Division of the Speech Communication Association.

    ERIC Educational Resources Information Center

    Friedrich, Gustav W.; And Others

    This paper focuses on performance-based teacher education in speech communication as an alternative to the traditional method of preparing speech teachers. Following an introductory argument stating the case for performance-based teacher education in speech communication, contents include: "Competency Based Teacher Training: A Perspective on a Set…

  5. Improved perception of speech in noise and Mandarin tones with acoustic simulations of harmonic coding for cochlear implants.

    PubMed

    Li, Xing; Nie, Kaibao; Imennov, Nikita S; Won, Jong Ho; Drennan, Ward R; Rubinstein, Jay T; Atlas, Les E

    2012-11-01

    Harmonic and temporal fine structure (TFS) information are important cues for speech perception in noise and music perception. However, due to the inherently coarse spectral and temporal resolution in electric hearing, the question of how to deliver harmonic and TFS information to cochlear implant (CI) users remains unresolved. A harmonic-single-sideband-encoder [(HSSE); Nie et al. (2008). Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing; Lie et al., (2010). Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing] strategy has been proposed that explicitly tracks the harmonics in speech and transforms them into modulators conveying both amplitude modulation and fundamental frequency information. For unvoiced speech, HSSE transforms the TFS into a slowly varying yet still noise-like signal. To investigate its potential, four- and eight-channel vocoder simulations of HSSE and the continuous-interleaved-sampling (CIS) strategy were implemented, respectively. Using these vocoders, five normal-hearing subjects' speech recognition performance was evaluated under different masking conditions; another five normal-hearing subjects' Mandarin tone identification performance was also evaluated. Additionally, the neural discharge patterns evoked by HSSE- and CIS-encoded Mandarin tone stimuli were simulated using an auditory nerve model. All subjects scored significantly higher with HSSE than with CIS vocoders. The modeling analysis demonstrated that HSSE can convey temporal pitch cues better than CIS. Overall, the results suggest that HSSE is a promising strategy to enhance speech perception with CIs. PMID:23145619

  6. Analysis of Acoustic Features in Speakers with Cognitive Disorders and Speech Impairments

    NASA Astrophysics Data System (ADS)

    Saz, Oscar; Simón, Javier; Rodríguez, W. Ricardo; Lleida, Eduardo; Vaquero, Carlos

    2009-12-01

    This work presents the results in the analysis of the acoustic features (formants and the three suprasegmental features: tone, intensity and duration) of the vowel production in a group of 14 young speakers suffering different kinds of speech impairments due to physical and cognitive disorders. A corpus with unimpaired children's speech is used to determine the reference values for these features in speakers without any kind of speech impairment within the same domain of the impaired speakers; this is 57 isolated words. The signal processing to extract the formant and pitch values is based on a Linear Prediction Coefficients (LPCs) analysis of the segments considered as vowels in a Hidden Markov Model (HMM) based Viterbi forced alignment. Intensity and duration are also based in the outcome of the automated segmentation. As main conclusion of the work, it is shown that intelligibility of the vowel production is lowered in impaired speakers even when the vowel is perceived as correct by human labelers. The decrease in intelligibility is due to a 30% of increase in confusability in the formants map, a reduction of 50% in the discriminative power in energy between stressed and unstressed vowels and to a 50% increase of the standard deviation in the length of the vowels. On the other hand, impaired speakers keep good control of tone in the production of stressed and unstressed vowels.

  7. Speech intelligibility in rooms: Effect of prior listening exposure interacts with room acoustics.

    PubMed

    Zahorik, Pavel; Brandewie, Eugene J

    2016-07-01

    There is now converging evidence that a brief period of prior listening exposure to a reverberant room can influence speech understanding in that environment. Although the effect appears to depend critically on the amplitude modulation characteristic of the speech signal reaching the ear, the extent to which the effect may be influenced by room acoustics has not been thoroughly evaluated. This study seeks to fill this gap in knowledge by testing the effect of prior listening exposure or listening context on speech understanding in five different simulated sound fields, ranging from anechoic space to a room with broadband reverberation time (T60) of approximately 3 s. Although substantial individual variability in the effect was observed and quantified, the context effect was, on average, strongly room dependent. At threshold, the effect was minimal in anechoic space, increased to a maximum of 3 dB on average in moderate reverberation (T60 = 1 s), and returned to minimal levels again in high reverberation. This interaction suggests that the functional effects of prior listening exposure may be limited to sound fields with moderate reverberation (0.4 ≤ T60 ≤ 1 s). PMID:27475133

  8. Communication and cooperation in underwater acoustic networks

    NASA Astrophysics Data System (ADS)

    Yerramalli, Srinivas

    In this thesis, we present a study of several problems related to underwater point to point communications and network formation. We explore techniques to improve the achievable data rate on a point to point link using better physical layer techniques and then study sensor cooperation which improves the throughput and reliability in an underwater network. Robust point-to-point communications in underwater networks has become increasingly critical in several military and civilian applications related to underwater communications. We present several physical layer signaling and detection techniques tailored to the underwater channel model to improve the reliability of data detection. First, a simplified underwater channel model in which the time scale distortion on each path is assumed to be the same (single scale channel model in contrast to a more general multi scale model). A novel technique, which exploits the nature of OFDM signaling and the time scale distortion, called Partial FFT Demodulation is derived. It is observed that this new technique has some unique interference suppression properties and performs better than traditional equalizers in several scenarios of interest. Next, we consider the multi scale model for the underwater channel and assume that single scale processing is performed at the receiver. We then derive optimized front end pre-processing techniques to reduce the interference caused during single scale processing of signals transmitted on a multi-scale channel. We then propose an improvised channel estimation technique using dictionary optimization methods for compressive sensing and show that significant performance gains can be obtained using this technique. In the next part of this thesis, we consider the problem of sensor node cooperation among rational nodes whose objective is to improve their individual data rates. We first consider the problem of transmitter cooperation in a multiple access channel and investigate the stability of

  9. Discrimination of Speech Stimuli Based on Neuronal Response Phase Patterns Depends on Acoustics But Not Comprehension

    PubMed Central

    Poeppel, David

    2010-01-01

    Speech stimuli give rise to neural activity in the listener that can be observed as waveforms using magnetoencephalography. Although waveforms vary greatly from trial to trial due to activity unrelated to the stimulus, it has been demonstrated that spoken sentences can be discriminated based on theta-band (3–7 Hz) phase patterns in single-trial response waveforms. Furthermore, manipulations of the speech signal envelope and fine structure that reduced intelligibility were found to produce correlated reductions in discrimination performance, suggesting a relationship between theta-band phase patterns and speech comprehension. This study investigates the nature of this relationship, hypothesizing that theta-band phase patterns primarily reflect cortical processing of low-frequency (<40 Hz) modulations present in the acoustic signal and required for intelligibility, rather than processing exclusively related to comprehension (e.g., lexical, syntactic, semantic). Using stimuli that are quite similar to normal spoken sentences in terms of low-frequency modulation characteristics but are unintelligible (i.e., their time-inverted counterparts), we find that discrimination performance based on theta-band phase patterns is equal for both types of stimuli. Consistent with earlier findings, we also observe that whereas theta-band phase patterns differ across stimuli, power patterns do not. We use a simulation model of the single-trial response to spoken sentence stimuli to demonstrate that phase-locked responses to low-frequency modulations of the acoustic signal can account not only for the phase but also for the power results. The simulation offers insight into the interpretation of the empirical results with respect to phase-resetting and power-enhancement models of the evoked response. PMID:20484530

  10. Communicating headings and preview sentences in text and speech.

    PubMed

    Lorch, Robert F; Chen, Hung-Tao; Lemarié, Julie

    2012-09-01

    Two experiments tested the effects of preview sentences and headings on the quality of college students' outlines of informational texts. Experiment 1 found that performance was much better in the preview sentences condition than in a no-signals condition for both printed text and text-to-speech (TTS) audio rendering of the printed text. In contrast, performance in the headings condition was good for the printed text but poor for the auditory presentation because the TTS software failed to communicate nonverbal information carried by the visual headings. Experiment 2 compared outlining performance for five headings conditions during TTS presentation. Using a theoretical framework, "signaling available, relevant, accessible" (SARA) information, to provide an analysis of the information content of headings in the printed text, the manipulation of the headings systematically restored information that was omitted by the TTS application in Experiment 1. The result was that outlining performance improved to levels similar to the visual headings condition of Experiment 1. It is argued that SARA is a useful framework for guiding future development of TTS software for a wide variety of text signaling devices, not just headings. PMID:22866682

  11. Brain mechanisms of acoustic communication in humans and nonhuman primates: an evolutionary perspective.

    PubMed

    Ackermann, Hermann; Hage, Steffen R; Ziegler, Wolfram

    2014-12-01

    Any account of "what is special about the human brain" (Passingham 2008) must specify the neural basis of our unique ability to produce speech and delineate how these remarkable motor capabilities could have emerged in our hominin ancestors. Clinical data suggest that the basal ganglia provide a platform for the integration of primate-general mechanisms of acoustic communication with the faculty of articulate speech in humans. Furthermore, neurobiological and paleoanthropological data point at a two-stage model of the phylogenetic evolution of this crucial prerequisite of spoken language: (i) monosynaptic refinement of the projections of motor cortex to the brainstem nuclei that steer laryngeal muscles, presumably, as part of a "phylogenetic trend" associated with increasing brain size during hominin evolution; (ii) subsequent vocal-laryngeal elaboration of cortico-basal ganglia circuitries, driven by human-specific FOXP2 mutations.;>This concept implies vocal continuity of spoken language evolution at the motor level, elucidating the deep entrenchment of articulate speech into a "nonverbal matrix" (Ingold 1994), which is not accounted for by gestural-origin theories. Moreover, it provides a solution to the question for the adaptive value of the "first word" (Bickerton 2009) since even the earliest and most simple verbal utterances must have increased the versatility of vocal displays afforded by the preceding elaboration of monosynaptic corticobulbar tracts, giving rise to enhanced social cooperation and prestige. At the ontogenetic level, the proposed model assumes age-dependent interactions between the basal ganglia and their cortical targets, similar to vocal learning in some songbirds. In this view, the emergence of articulate speech builds on the "renaissance" of an ancient organizational principle and, hence, may represent an example of "evolutionary tinkering" (Jacob 1977).

  12. A supervisory control policy over an acoustic communication network

    NASA Astrophysics Data System (ADS)

    Farhadi, Alireza; Dumon, Jonathan; Canudas-de-Wit, Carlos

    2015-05-01

    This paper presents a supervisory multi-agent control policy over an acoustic communication network subject to imperfections (packet dropout and transmission delay) for localisation of an underwater flow source (e.g., source of chemical pollution, fresh water, etc.) with an unknown location at the bottom of the ocean. A two-loop control policy combined with a coding strategy for reliable communication is presented to perform the above task. A simulator is developed and used to evaluate the trade-offs between quality of communication, transmission delay and control for a fleet of autonomous underwater vehicles supervised over a noisy acoustic communication network by an autonomous surface vessel. It is illustrated that without compensation of the effects of severe random packet dropout, localisation of an unknown underwater flow source is not possible for the condition simulated just by implementing a two-loop control policy. But a two-loop control policy combined with a strategy for reliable communication locates the unknown location of flow source.

  13. Doppler compensated underwater acoustic communication system

    NASA Astrophysics Data System (ADS)

    Raj, Anand; George, Binu; Supiya, M. H.; Kurian, James; Pillai, P. R. Saseendran

    2001-05-01

    Spread spectrum methods are used in communication systems to provide a low probability of intercept in hostile environments and multiple access capability in systems shared by many users as well as to provide high processing gain in channels where the transmitted signal is distorted by multipath effects. Such systems serve to be an effective tool for underwater telemetry environments, where multipath propagation effect and Doppler spreading is seen to be more predominant. This paper describes the implementation of a Doppler compensated underwater telemetry system based on CDMA technique. The system consists of multiple CDMA transmitters and a phase locked loop based carrier recoverable CDMA receiver. The effects of the Doppler shift can be compensated by the carrier recovery subsystem in the demodulator, based on PLL technique, which extracts the carrier frequency/phase and simultaneously demodulates the signal. The decision device in the receiver consists of a PN sequence generator as well as a bank of correlators, which are used to determine the data transmitted. The system simulation has been implemented in MATLAB. The advantage of this system is that multiple transmitting stations can transmit simultaneously to a central receiver, thereby increasing the system throughput.

  14. An acoustical assessment of pitch-matching accuracy in relation to speech frequency, speech frequency range, age and gender in preschool children

    NASA Astrophysics Data System (ADS)

    Trollinger, Valerie L.

    This study investigated the relationship between acoustical measurement of singing accuracy in relationship to speech fundamental frequency, speech fundamental frequency range, age and gender in preschool-aged children. Seventy subjects from Southeastern Pennsylvania; the San Francisco Bay Area, California; and Terre Haute, Indiana, participated in the study. Speech frequency was measured by having the subjects participate in spontaneous and guided speech activities with the researcher, with 18 diverse samples extracted from each subject's recording for acoustical analysis for fundamental frequency in Hz with the CSpeech computer program. The fundamental frequencies were averaged together to derive a mean speech frequency score for each subject. Speech range was calculated by subtracting the lowest fundamental frequency produced from the highest fundamental frequency produced, resulting in a speech range measured in increments of Hz. Singing accuracy was measured by having the subjects each echo-sing six randomized patterns using the pitches Middle C, D, E, F♯, G and A (440), using the solfege syllables of Do and Re, which were recorded by a 5-year-old female model. For each subject, 18 samples of singing were recorded. All samples were analyzed by the CSpeech for fundamental frequency. For each subject, deviation scores in Hz were derived by calculating the difference between what the model sang in Hz and what the subject sang in response in Hz. Individual scores for each child consisted of an overall mean total deviation frequency, mean frequency deviations for each pattern, and mean frequency deviation for each pitch. Pearson correlations, MANOVA and ANOVA analyses, Multiple Regressions and Discriminant Analysis revealed the following findings: (1) moderate but significant (p < .001) relationships emerged between mean speech frequency and the ability to sing the pitches E, F♯, G and A in the study; (2) mean speech frequency also emerged as the strongest

  15. Acoustic diagnosis of pulmonary hypertension: automated speech- recognition-inspired classification algorithm outperforms physicians.

    PubMed

    Kaddoura, Tarek; Vadlamudi, Karunakar; Kumar, Shine; Bobhate, Prashant; Guo, Long; Jain, Shreepal; Elgendi, Mohamed; Coe, James Y; Kim, Daniel; Taylor, Dylan; Tymchak, Wayne; Schuurmans, Dale; Zemp, Roger J; Adatia, Ian

    2016-01-01

    We hypothesized that an automated speech- recognition-inspired classification algorithm could differentiate between the heart sounds in subjects with and without pulmonary hypertension (PH) and outperform physicians. Heart sounds, electrocardiograms, and mean pulmonary artery pressures (mPAp) were recorded simultaneously. Heart sound recordings were digitized to train and test speech-recognition-inspired classification algorithms. We used mel-frequency cepstral coefficients to extract features from the heart sounds. Gaussian-mixture models classified the features as PH (mPAp ≥ 25 mmHg) or normal (mPAp < 25 mmHg). Physicians blinded to patient data listened to the same heart sound recordings and attempted a diagnosis. We studied 164 subjects: 86 with mPAp ≥ 25 mmHg (mPAp 41 ± 12 mmHg) and 78 with mPAp < 25 mmHg (mPAp 17 ± 5 mmHg) (p  < 0.005). The correct diagnostic rate of the automated speech-recognition-inspired algorithm was 74% compared to 56% by physicians (p = 0.005). The false positive rate for the algorithm was 34% versus 50% (p = 0.04) for clinicians. The false negative rate for the algorithm was 23% and 68% (p = 0.0002) for physicians. We developed an automated speech-recognition-inspired classification algorithm for the acoustic diagnosis of PH that outperforms physicians that could be used to screen for PH and encourage earlier specialist referral. PMID:27609672

  16. Acoustic diagnosis of pulmonary hypertension: automated speech- recognition-inspired classification algorithm outperforms physicians

    PubMed Central

    Kaddoura, Tarek; Vadlamudi, Karunakar; Kumar, Shine; Bobhate, Prashant; Guo, Long; Jain, Shreepal; Elgendi, Mohamed; Coe, James Y; Kim, Daniel; Taylor, Dylan; Tymchak, Wayne; Schuurmans, Dale; Zemp, Roger J.; Adatia, Ian

    2016-01-01

    We hypothesized that an automated speech- recognition-inspired classification algorithm could differentiate between the heart sounds in subjects with and without pulmonary hypertension (PH) and outperform physicians. Heart sounds, electrocardiograms, and mean pulmonary artery pressures (mPAp) were recorded simultaneously. Heart sound recordings were digitized to train and test speech-recognition-inspired classification algorithms. We used mel-frequency cepstral coefficients to extract features from the heart sounds. Gaussian-mixture models classified the features as PH (mPAp ≥ 25 mmHg) or normal (mPAp < 25 mmHg). Physicians blinded to patient data listened to the same heart sound recordings and attempted a diagnosis. We studied 164 subjects: 86 with mPAp ≥ 25 mmHg (mPAp 41 ± 12 mmHg) and 78 with mPAp < 25 mmHg (mPAp 17 ± 5 mmHg) (p  < 0.005). The correct diagnostic rate of the automated speech-recognition-inspired algorithm was 74% compared to 56% by physicians (p = 0.005). The false positive rate for the algorithm was 34% versus 50% (p = 0.04) for clinicians. The false negative rate for the algorithm was 23% and 68% (p = 0.0002) for physicians. We developed an automated speech-recognition-inspired classification algorithm for the acoustic diagnosis of PH that outperforms physicians that could be used to screen for PH and encourage earlier specialist referral. PMID:27609672

  17. Acoustic diagnosis of pulmonary hypertension: automated speech- recognition-inspired classification algorithm outperforms physicians

    NASA Astrophysics Data System (ADS)

    Kaddoura, Tarek; Vadlamudi, Karunakar; Kumar, Shine; Bobhate, Prashant; Guo, Long; Jain, Shreepal; Elgendi, Mohamed; Coe, James Y.; Kim, Daniel; Taylor, Dylan; Tymchak, Wayne; Schuurmans, Dale; Zemp, Roger J.; Adatia, Ian

    2016-09-01

    We hypothesized that an automated speech- recognition-inspired classification algorithm could differentiate between the heart sounds in subjects with and without pulmonary hypertension (PH) and outperform physicians. Heart sounds, electrocardiograms, and mean pulmonary artery pressures (mPAp) were recorded simultaneously. Heart sound recordings were digitized to train and test speech-recognition-inspired classification algorithms. We used mel-frequency cepstral coefficients to extract features from the heart sounds. Gaussian-mixture models classified the features as PH (mPAp ≥ 25 mmHg) or normal (mPAp < 25 mmHg). Physicians blinded to patient data listened to the same heart sound recordings and attempted a diagnosis. We studied 164 subjects: 86 with mPAp ≥ 25 mmHg (mPAp 41 ± 12 mmHg) and 78 with mPAp < 25 mmHg (mPAp 17 ± 5 mmHg) (p  < 0.005). The correct diagnostic rate of the automated speech-recognition-inspired algorithm was 74% compared to 56% by physicians (p = 0.005). The false positive rate for the algorithm was 34% versus 50% (p = 0.04) for clinicians. The false negative rate for the algorithm was 23% and 68% (p = 0.0002) for physicians. We developed an automated speech-recognition-inspired classification algorithm for the acoustic diagnosis of PH that outperforms physicians that could be used to screen for PH and encourage earlier specialist referral.

  18. Acoustical and Intelligibility Test of the Vocera(Copyright) B3000 Communication Badge

    NASA Technical Reports Server (NTRS)

    Archer, Ronald; Litaker, Harry; Chu, Shao-Sheng R.; Simon, Cory; Romero, Andy; Moses, Haifa

    2012-01-01

    To communicate with each other or ground support, crew members on board the International Space Station (ISS) currently use the Audio Terminal Units (ATU), which are located in each ISS module. However, to use the ATU, crew members must stop their current activity, travel to a panel, and speak into a wall-mounted microphone, or use either a handheld microphone or a Crew Communication Headset that is connected to a panel. These actions unnecessarily may increase task times, lower productivity, create cable management issues, and thus increase crew frustration. Therefore, the Habitability and Human Factors and Human Interface Branches at the NASA Johnson Space Center (JSC) are currently investigating a commercial-off-the-shelf (COTS) wireless communication system, Vocera(C), as a near-term solution for ISS communication. The objectives of the acoustics and intelligibility testing of this system were to answer the following questions: 1. How intelligibly can a human hear the transmitted message from a Vocera(c) badge in three different noise environments (Baseline = 20 dB, US Lab Module = 58 dB, Russian Module = 70.6 dB)? 2. How accurate is the Vocera(C) badge at recognizing voice commands in three different noise environments? 3. What body location (chest, upper arm, or shoulder) is optimal for speech intelligibility and voice recognition accuracy of the Vocera(C) badge on a human in three different noise environments?

  19. Transient Auditory Storage of Acoustic Details Is Associated with Release of Speech from Informational Masking in Reverberant Conditions

    ERIC Educational Resources Information Center

    Huang, Ying; Huang, Qiang; Chen, Xun; Wu, Xihong; Li, Liang

    2009-01-01

    Perceptual integration of the sound directly emanating from the source with reflections needs both temporal storage and correlation computation of acoustic details. We examined whether the temporal storage is frequency dependent and associated with speech unmasking. In Experiment 1, a break in correlation (BIC) between interaurally correlated…

  20. Discrimination and Comprehension of Synthetic Speech by Students with Visual Impairments: The Case of Similar Acoustic Patterns

    ERIC Educational Resources Information Center

    Papadopoulos, Konstantinos; Argyropoulos, Vassilios S.; Kouroupetroglou, Georgios

    2008-01-01

    This study examined the perceptions held by sighted students and students with visual impairments of the intelligibility and comprehensibility of similar acoustic patterns produced by synthetic speech. It determined the types of errors the students made and compared the performance of the two groups on auditory discrimination and comprehension.

  1. Acoustic patterns of infant vocalizations expressing emotions and communicative functions.

    PubMed

    Papaeliou, C; Minadakis, G; Cavouras, D

    2002-04-01

    The present study aimed at identifying the acoustic pattern of vocalizations, produced by 7- to 11-month-old infants, that were interpreted by their mothers as expressing emotions or communicative functions. Participants were 6 healthy, first-born English infants, 3 boys and 3 girls, and their mothers. The acoustic analysis of the vocalizations was performed using a pattern recognition (PR) software system. A PR system not only calculates signal features, it also automatically detects patterns in the arrangement of such features. The following results were obtained: (a) the PR system distinguished vocalizations interpreted as emotions from vocalizations interpreted as communicative functions with an overall accuracy of 87.34%; (b) the classification accuracy of the PR system for vocalizations that convey emotions was 85.4% and for vocalizations that convey communicative functions was 89.5%; and (c) compared to vocalizations that express emotions, vocalizations that express communicative functions were shorter, displayed lower fundamental frequency values, and had greater overall intensity. These findings suggest that in the second half of the first year, infants possess a vocal repertoire that contributes to regulating cooperative interaction with their mothers, which is considered one of the major prerequisites for language acquisition.

  2. Communication calls of Japanese pipistrelle bats: Examination by acoustic characteristics

    NASA Astrophysics Data System (ADS)

    Osawa, Yoshiki; Kondo, Yoko; Nagato, Tsuyoshi; Watanabe, Yoshiaki; Riquimaroux, Hiroshi

    2005-09-01

    We classified communication calls of Japanese pipistrelle bats (Pipistrellus abramus) by acoustic characteristics D. The Japanese pipistrelles emitted communication calls that were completely different from FM echolocation calls. Data showed that in general duration of communication calls was longer than that of echolocation calls (e.g., echolocation call, 1 ms; long CF communication call, 50 ms) and that frequency of commu-nication calls were lower than that of echolocation calls (e.g., echolocation call, 80-40 kHz; long CF communication call, about 14 kHz). Typical classified communication calls were as follows: slightly short CF call (20 ms, 14 kHz), short CF call (5 ms, 14 kHz), slightly long CF call (30 ms, 14 kHz), long CF call (50 ms, 14 kHz), slightly long FM call (15 ms, 30-15 kHz), long FM call (20 ms, 25-15 kHz), complicated FM call (10-50 ms, 25-15 kHz), short intermittent CF call (1 ms, 14 kHz) and noise call (20-100 ms, below 60 kHz). Details will be discussed more specifically. [Research supported by a grant to RCAST at Doshisha Univ. from MEXT and by the Innovative Cluster Creation Project promoted by MEXT.

  3. Measuring Communicative Participation: A Review of Self-Report Instruments in Speech-Language Pathology

    PubMed Central

    Eadie, Tanya L.; Yorkston, Kathryn M.; Klasner, Estelle R.; Dudgeon, Brian J.; Deitz, Jean C.; Baylor, Carolyn R.; Miller, Robert M.; Amtmann, Dagmar

    2009-01-01

    Purpose To assess the adequacy of self-report instruments in speech-language pathology for measuring a construct called communicative participation. Method Six instruments were evaluated relative to (a) the construct measured, (b) the relevance of individual items to communicative participation, and (c) their psychometric properties. Results No instrument exclusively measured communicative participation. Twenty-six percent (n = 34) of all items (N = 132) across the reviewed instruments were consistent with communicative participation. The majority (76%) of the 34 items were associated with general communication, while the remaining 24% of the items were associated with communication at work, during leisure, or for establishing relationships. Instruments varied relative to psychometric properties. Conclusions No existing self-report instruments in speech-language pathology were found to be solely dedicated to measuring communicative participation. Developing an instrument for measuring communicative participation is essential for meeting the requirements of our scope of practice. PMID:17102143

  4. Energy scavenging system by acoustic wave and integrated wireless communication

    NASA Astrophysics Data System (ADS)

    Kim, Albert

    The purpose of the project was developing an energy-scavenging device for other bio implantable devices. Researchers and scientist have studied energy scavenging method because of the limitation of traditional power source, especially for bio-implantable devices. In this research, piezoelectric power generator that activates by acoustic wave, or music was developed. Follow by power generator, a wireless communication also integrated with the device for monitoring the power generation. The Lead Zirconate Titanate (PZT) bimorph cantilever with a proof mass at the free end tip was studied to convert acoustic wave to power. The music or acoustic wave played through a speaker to vibrate piezoelectric power generator. The LC circuit integrated with the piezoelectric material for purpose of wireless monitoring power generation. However, wireless monitoring can be used as wireless power transmission, which means the signal received via wireless communication also can be used for power for other devices. Size of 74 by 7 by 7cm device could generate and transmit 100mVp from 70 mm distance away with electrical resonant frequency at 420.2 kHz..

  5. Comparing the effects of reverberation and of noise on speech recognition in simulated electric-acoustic listening

    PubMed Central

    Helms Tillery, Kate; Brown, Christopher A.; Bacon, Sid P.

    2012-01-01

    Cochlear implant users report difficulty understanding speech in both noisy and reverberant environments. Electric-acoustic stimulation (EAS) is known to improve speech intelligibility in noise. However, little is known about the potential benefits of EAS in reverberation, or about how such benefits relate to those observed in noise. The present study used EAS simulations to examine these questions. Sentences were convolved with impulse responses from a model of a room whose estimated reverberation times were varied from 0 to 1 sec. These reverberated stimuli were then vocoded to simulate electric stimulation, or presented as a combination of vocoder plus low-pass filtered speech to simulate EAS. Monaural sentence recognition scores were measured in two conditions: reverberated speech and speech in a reverberated noise. The long-term spectrum and amplitude modulations of the noise were equated to the reverberant energy, allowing a comparison of the effects of the interferer (speech vs noise). Results indicate that, at least in simulation, (1) EAS provides significant benefit in reverberation; (2) the benefits of EAS in reverberation may be underestimated by those in a comparable noise; and (3) the EAS benefit in reverberation likely arises from partially preserved cues in this background accessible via the low-frequency acoustic component. PMID:22280603

  6. Acoustics in educational settings. Subcommittee on Acoustics in Educational Settings of the Bioacoustics Standards and Noise Standards Committee American Speech-Language-Hearing Association.

    PubMed

    1995-03-01

    The Americans with Disabilities Act (enacted July 26, 1990) has brought into focus the need for removing barriers and improving accessibility of all buildings and facilities. It is clear that the definition of barrier must be expanded to include not only structural features that limit physical accessibility, but also acoustical barriers that limit access to communication and information. Acoustical interference caused by inappropriate levels of background noise and reverberation presents a barrier to learning and communication in educational settings and school-sponsored extracurricular activities, particularly for students with hearing loss or other language/learning concerns. ASHA has provided these guidelines and acoustical improvement strategies in order to assist communication-related professionals, teachers, school officials, architects, contractors, state education agencies, and others in developing the best possible learning environment for all students. Additional research on both the acoustical characteristics of learning environments and the communication requirements of learners is encouraged. PMID:7696882

  7. Exposure to seismic survey alters blue whale acoustic communication.

    PubMed

    Di Iorio, Lucia; Clark, Christopher W

    2010-02-23

    The ability to perceive biologically important sounds is critical to marine mammals, and acoustic disturbance through human-generated noise can interfere with their natural functions. Sounds from seismic surveys are intense and have peak frequency bands overlapping those used by baleen whales, but evidence of interference with baleen whale acoustic communication is sparse. Here we investigated whether blue whales (Balaenoptera musculus) changed their vocal behaviour during a seismic survey that deployed a low-medium power technology (sparker). We found that blue whales called consistently more on seismic exploration days than on non-exploration days as well as during periods within a seismic survey day when the sparker was operating. This increase was observed for the discrete, audible calls that are emitted during social encounters and feeding. This response presumably represents a compensatory behaviour to the elevated ambient noise from seismic survey operations. PMID:19776059

  8. Impact of source depth on coherent underwater acoustic communications.

    PubMed

    Song, Aijun; Badiey, Mohsen; Song, H C; Hodgkiss, W S

    2010-08-01

    A recent paper [Song et al., J. Acoust. Soc. Am. 123, 856-865 (2008)] investigated ocean variability impact on coherent underwater acoustic communications (8-16 kHz) for a single near-seafloor transmitter in shallow water during an extended period (27 h). This letter extends that investigation to various source depths and receiver subarrays. Specifically, the middle water column source, which is either in or out of the thermocline, experiences performance variability of 6-7 dB in terms of output signal-to-noise ratio. Further, the source below the thermocline consistently outperforms the source above the thermocline when the receiver subarray is located below the thermocline.

  9. Exposure to seismic survey alters blue whale acoustic communication.

    PubMed

    Di Iorio, Lucia; Clark, Christopher W

    2010-02-23

    The ability to perceive biologically important sounds is critical to marine mammals, and acoustic disturbance through human-generated noise can interfere with their natural functions. Sounds from seismic surveys are intense and have peak frequency bands overlapping those used by baleen whales, but evidence of interference with baleen whale acoustic communication is sparse. Here we investigated whether blue whales (Balaenoptera musculus) changed their vocal behaviour during a seismic survey that deployed a low-medium power technology (sparker). We found that blue whales called consistently more on seismic exploration days than on non-exploration days as well as during periods within a seismic survey day when the sparker was operating. This increase was observed for the discrete, audible calls that are emitted during social encounters and feeding. This response presumably represents a compensatory behaviour to the elevated ambient noise from seismic survey operations.

  10. A Dual Communication and Imaging Underwater Acoustic System

    NASA Astrophysics Data System (ADS)

    Fu, Tricia C.

    A dual communication and imaging underwater acoustic system is proposed and developed throughout this dissertation. Due to the wide variation in underwater channel characteristics, the research here focuses more on robustness to multipath in the shallow underwater acoustic environment, rather than high bit-rate applications and signaling schemes. Lower bit-rate (in the hundreds of bits per second (bps) to low kbps), applications such as the transfer of ecological telemetry data, e.g. conductivity or temperature data, are the primary focus of this dissertation. The parallels between direct sequence spread spectrum in digital communication and pulse-echo with pulse compression in imaging, and channel estimation in communication and range profile estimation in imaging are drawn, leading to a unified communications and imaging platform. A digital communication algorithm for channel order and channel coefficient estimation and symbol demodulation using Matching Pursuit (MP) with Generalized Multiple Hypothesis Testing (GMHT) is implemented in programmable DSP in real time with field experiment results in varying underwater environments for the single receiver (Rx), single transmitter (Tx) case. The custom and off-the-shelf hardware used in the single receiver, single transmitter set of experiments are detailed as well. This work is then extended to the single-input multiple-output (SIMO) case, and then to the full multiple-input multiple-output (MIMO) case. The results of channel estimation are used for simple range profile imaging reconstructions. Successful simulated and experimental results for both transducer array configurations are presented and analyzed. Non-real-time symbol demodulation and channel estimation is performed using experimental data from a scaled testing environment. New hardware based on cost-effective fish-finder transducers for a 6 Rx--1 Tx and 6 Rx--4 Tx transducer array is detailed. Lastly, in an application that is neither communication nor

  11. Recognition of emotions in Mexican Spanish speech: an approach based on acoustic modelling of emotion-specific vowels.

    PubMed

    Caballero-Morales, Santiago-Omar

    2013-01-01

    An approach for the recognition of emotions in speech is presented. The target language is Mexican Spanish, and for this purpose a speech database was created. The approach consists in the phoneme acoustic modelling of emotion-specific vowels. For this, a standard phoneme-based Automatic Speech Recognition (ASR) system was built with Hidden Markov Models (HMMs), where different phoneme HMMs were built for the consonants and emotion-specific vowels associated with four emotional states (anger, happiness, neutral, sadness). Then, estimation of the emotional state from a spoken sentence is performed by counting the number of emotion-specific vowels found in the ASR's output for the sentence. With this approach, accuracy of 87-100% was achieved for the recognition of emotional state of Mexican Spanish speech.

  12. Recognition of Emotions in Mexican Spanish Speech: An Approach Based on Acoustic Modelling of Emotion-Specific Vowels

    PubMed Central

    Caballero-Morales, Santiago-Omar

    2013-01-01

    An approach for the recognition of emotions in speech is presented. The target language is Mexican Spanish, and for this purpose a speech database was created. The approach consists in the phoneme acoustic modelling of emotion-specific vowels. For this, a standard phoneme-based Automatic Speech Recognition (ASR) system was built with Hidden Markov Models (HMMs), where different phoneme HMMs were built for the consonants and emotion-specific vowels associated with four emotional states (anger, happiness, neutral, sadness). Then, estimation of the emotional state from a spoken sentence is performed by counting the number of emotion-specific vowels found in the ASR's output for the sentence. With this approach, accuracy of 87–100% was achieved for the recognition of emotional state of Mexican Spanish speech. PMID:23935410

  13. Repetition and Emotive Communication in Music Versus Speech

    PubMed Central

    Margulis, Elizabeth Hellmuth

    2013-01-01

    Music and speech are often placed alongside one another as comparative cases. Their relative overlaps and disassociations have been well explored (e.g., Patel, 2008). But one key attribute distinguishing these two domains has often been overlooked: the greater preponderance of repetition in music in comparison to speech. Recent fMRI studies have shown that familiarity – achieved through repetition – is a critical component of emotional engagement with music (Pereira et al., 2011). If repetition is fundamental to emotional responses to music, and repetition is a key distinguisher between the domains of music and speech, then close examination of the phenomenon of repetition might help clarify the ways that music elicits emotion differently than speech. PMID:23576998

  14. Children with Autistic Spectrum Disorders and Speech-Generating Devices: Communication in Different Activities at Home

    ERIC Educational Resources Information Center

    Thunberg, Gunilla; Ahlsen, Elisabeth; Sandberg, Annika Dahlgren

    2007-01-01

    The communication of four children with autistic spectrum disorder was investigated when they were supplied with a speech-generating device (SGD) in three different activities in their home environment: mealtime, story reading and "sharing experiences of the preschool day". An activity based communication analysis, in which collective and…

  15. The Status of Ethics Scholarship in Speech Communication Journals from 1915 to 1985.

    ERIC Educational Resources Information Center

    Arnett, Ronald C.

    To examine the theoretical status of ethics scholarship and to explore the historical and present directions of ethics in human communication research, this paper reviews more than 100 articles drawn from the speech communication literature. Following a brief introduction that sets forth the criteria for article selection, the paper discusses…

  16. The Experimental Social Scientific Model in Speech Communication Research: Influences and Consequences.

    ERIC Educational Resources Information Center

    Ferris, Sharmila Pixy

    A substantial number of published articles in speech communication research today is experimental/social scientific in nature. It is only in the past decade that scholars have begun to put the history of communication under the lens. Early advocates of the adoption of the method of social scientific inquiry were J. A. Winans, J. M. O'Neill, and C.…

  17. Acoustic markers of prominence influence infants' and adults' segmentation of speech sequences.

    PubMed

    Bion, Ricardo A H; Benavides-Varela, Silvia; Nespor, Marina

    2011-03-01

    Two experiments investigated the way acoustic markers of prominence influence the grouping of speech sequences by adults and 7-month-old infants. In the first experiment, adults were familiarized with and asked to memorize sequences of adjacent syllables that alternated in either pitch or duration. During the test phase, participants heard pairs of syllables with constant pitch and duration and were asked whether the syllables had appeared adjacently during familiarization. Adults were better at remembering pairs of syllables that during familiarization had short syllables preceding long syllables, or high-pitched syllables preceding low-pitched syllables. In the second experiment, infants were familiarized and tested with similar stimuli as in the first experiment, and their preference for pairs of syllables was accessed using the head-turn preference paradigm.When familiarized with syllables alternating in pitch, infants showed a preference to listen to pairs of syllables that had high pitch in the first syllable. However, no preference was found when the familiarization stream alternated in duration. It is proposed that these perceptual biases help infants and adults find linguistic units in the continuous speech stream.While the bias for grouping based on pitch appears early in development, biases for durational grouping might rely on more extensive linguistic experience. PMID:21524015

  18. Statistical evidence that musical universals derive from the acoustic characteristics of human speech

    NASA Astrophysics Data System (ADS)

    Schwartz, David; Howe, Catharine; Purves, Dale

    2003-04-01

    Listeners of all ages and societies produce a similar consonance ordering of chromatic scale tone combinations. Despite intense interest in this perceptual phenomenon over several millennia, it has no generally accepted explanation in physical, psychological, or physiological terms. Here we show that the musical universal of consonance ordering can be understood in terms of the statistical relationship between a pattern of sound pressure at the ear and the possible generative sources of the acoustic energy pattern. Since human speech is the principal naturally occurring source of tone-evoking (i.e., periodic) sound energy for human listeners, we obtained normalized spectra from more than 100000 recorded speech segments. The probability distribution of amplitude/frequency combinations derived from these spectra predicts both the fundamental frequency ratios that define the chromatic scale intervals and the consonance ordering of chromatic scale tone combinations. We suggest that these observations reveal the statistical character of the perceptual process by which the auditory system guides biologically successful behavior in response to inherently ambiguous sound stimuli.

  19. A machine for neural computation of acoustical patterns with application to real time speech recognition

    NASA Astrophysics Data System (ADS)

    Mueller, P.; Lazzaro, J.

    1986-08-01

    400 analog electronic neurons have been assembled and connected for the analysis and recognition of acoustical patterns, including speech. Input to the net comes from a set of 18 band pass filters (Qmax 300 dB/octave; 180 to 6000 Hz, log scale). The net is organized into two parts, the first performs in real time the decomposition of the input patterns into their primitives of energy, space (frequency) and time relations. The other part decodes the set of primitives. 216 neurons are dedicated to pattern decomposition. The output of the individual filters is rectified and fed to two sets of 18 neurons in an opponent center-surround organization of synaptic connections (``on center'' and (``off center''). These units compute maxima and minima of energy at different frequencies. The next two sets of neutrons compute the temporal boundaries (``on'') and ``off'') and the following two the movement of the energy maxima (formants) up or down the frequency axis. There are in addition ``hyperacuity'' units which expand the frequency resolution to 36, other units tuned to a particular range of duration of the ``on center'' units and others tuned exclusively to very low energy sounds. In order to recognize speech sounds at the phoneme or diphone level, the set of primitives belonging to the phoneme is decoded such that only one neuron or a non-overlapping group of neurons fire when the sound pattern is present at the input. For display and translation into phonetic symbols the output from these neurons is fed into an EPROM decoder and computer which displays in real time a phonetic representation of the speech input.

  20. Comparison study of time reversal OFDM acoustic communication with vector and scalar sensors

    NASA Astrophysics Data System (ADS)

    Wang, Zhongkang; Zhang, Hongtao; Xie, Zhe

    2012-11-01

    To compare the performance of time reversal orthogonal frequency division multiplexing (OFDM) acoustic communication on vector and scalar sensors, the vector and scalar acoustic fields were modeled. Time reversal OFDM acoustic communication was then simulated for each sensor type. These results are compared with data from the CAPEx'09 experiment. The abilityof particle velocity channels to achieve reliable acoustic communication, as predicted by the model, is confirmed with the experiment data. Experimental results show that vector receivers can reduce the required array size, in comparisonto hydrophone arrays, whileproviding comparable communication performance.

  1. A Fibre Bragg Grating Sensor as a Receiver for Acoustic Communications Signals

    PubMed Central

    Wild, Graham; Hinckley, Steven

    2011-01-01

    A Fibre Bragg Grating (FBG) acoustic sensor is used as a receiver for acoustic communications signals. Acoustic transmissions were generated in aluminium and Carbon Fibre Composite (CFC) panels. The FBG receiver was coupled to the bottom surface opposite a piezoelectric transmitter. For the CFC, a second FBG was embedded within the layup for comparison. We show the transfer function, frequency response, and transient response of the acoustic communications channels. In addition, the FBG receiver was used to detect Phase Shift Keying (PSK) communications signals, which was shown to be the most robust method in a highly resonant communications channel. PMID:22346585

  2. The effect of different open plan and enclosed classroom acoustic conditions on speech perception in Kindergarten children.

    PubMed

    Mealings, Kiri T; Demuth, Katherine; Buchholz, Jörg M; Dillon, Harvey

    2015-10-01

    Open plan classrooms, where several classes are in the same room, have recently re-emerged in Australian primary schools. This paper explores how the acoustics of four Kindergarten classrooms [an enclosed classroom (25 children), double classroom (44 children), fully open plan triple classroom (91 children), and a semi-open plan K-6 "21st century learning space" (205 children)] affect speech perception. Twenty-two to 23 5-6-year-old children in each classroom participated in an online four-picture choice speech perception test while adjacent classes engaged in quiet versus noisy activities. The noise levels recorded during the test were higher the larger the classroom, except in the noisy condition for the K-6 classroom, possibly due to acoustic treatments. Linear mixed effects models revealed children's performance accuracy and speed decreased as noise level increased. Additionally, children's speech perception abilities decreased the further away they were seated from the loudspeaker in noise levels above 50 dBA. These results suggest that fully open plan classrooms are not appropriate learning environments for critical listening activities with young children due to their high intrusive noise levels which negatively affect speech perception. If open plan classrooms are desired, they need to be acoustically designed to be appropriate for critical listening activities.

  3. Speaker height estimation from speech: Fusing spectral regression and statistical acoustic models.

    PubMed

    Hansen, John H L; Williams, Keri; Bořil, Hynek

    2015-08-01

    Estimating speaker height can assist in voice forensic analysis and provide additional side knowledge to benefit automatic speaker identification or acoustic model selection for automatic speech recognition. In this study, a statistical approach to height estimation that incorporates acoustic models within a non-uniform height bin width Gaussian mixture model structure as well as a formant analysis approach that employs linear regression on selected phones are presented. The accuracy and trade-offs of these systems are explored by examining the consistency of the results, location, and causes of error as well a combined fusion of the two systems using data from the TIMIT corpus. Open set testing is also presented using the Multi-session Audio Research Project corpus and publicly available YouTube audio to examine the effect of channel mismatch between training and testing data and provide a realistic open domain testing scenario. The proposed algorithms achieve a highly competitive performance to previously published literature. Although the different data partitioning in the literature and this study may prevent performance comparisons in absolute terms, the mean average error of 4.89 cm for males and 4.55 cm for females provided by the proposed algorithm on TIMIT utterances containing selected phones suggest a considerable estimation error decrease compared to past efforts.

  4. Speech timing and linguistic rhythm: on the acoustic bases of rhythm typologies.

    PubMed

    Rathcke, Tamara V; Smith, Rachel H

    2015-05-01

    Research into linguistic rhythm has been dominated by the idea that languages can be classified according to rhythmic templates, amenable to assessment by acoustic measures of vowel and consonant durations. This study tested predictions of two proposals explaining the bases of rhythmic typologies: the Rhythm Class Hypothesis which assumes that the templates arise from an extensive vs a limited use of durational contrasts, and the Control and Compensation Hypothesis which proposes that the templates are rooted in more vs less flexible speech production strategies. Temporal properties of segments, syllables and rhythmic feet were examined in two accents of British English, a "stress-timed" variety from Leeds, and a "syllable-timed" variety spoken by Panjabi-English bilinguals from Bradford. Rhythm metrics were calculated. A perception study confirmed that the speakers of the two varieties differed in their perceived rhythm. The results revealed that both typologies were informative in that to a certain degree, they predicted temporal patterns of the two varieties. None of the metrics tested was capable of adequately reflecting the temporal complexity found in the durational data. These findings contribute to the critical evaluation of the explanatory adequacy of rhythm metrics. Acoustic bases and limitations of the traditional rhythmic typologies are discussed.

  5. Speaker height estimation from speech: Fusing spectral regression and statistical acoustic models.

    PubMed

    Hansen, John H L; Williams, Keri; Bořil, Hynek

    2015-08-01

    Estimating speaker height can assist in voice forensic analysis and provide additional side knowledge to benefit automatic speaker identification or acoustic model selection for automatic speech recognition. In this study, a statistical approach to height estimation that incorporates acoustic models within a non-uniform height bin width Gaussian mixture model structure as well as a formant analysis approach that employs linear regression on selected phones are presented. The accuracy and trade-offs of these systems are explored by examining the consistency of the results, location, and causes of error as well a combined fusion of the two systems using data from the TIMIT corpus. Open set testing is also presented using the Multi-session Audio Research Project corpus and publicly available YouTube audio to examine the effect of channel mismatch between training and testing data and provide a realistic open domain testing scenario. The proposed algorithms achieve a highly competitive performance to previously published literature. Although the different data partitioning in the literature and this study may prevent performance comparisons in absolute terms, the mean average error of 4.89 cm for males and 4.55 cm for females provided by the proposed algorithm on TIMIT utterances containing selected phones suggest a considerable estimation error decrease compared to past efforts. PMID:26328721

  6. A voice-input voice-output communication aid for people with severe speech impairment.

    PubMed

    Hawley, Mark S; Cunningham, Stuart P; Green, Phil D; Enderby, Pam; Palmer, Rebecca; Sehgal, Siddharth; O'Neill, Peter

    2013-01-01

    A new form of augmentative and alternative communication (AAC) device for people with severe speech impairment-the voice-input voice-output communication aid (VIVOCA)-is described. The VIVOCA recognizes the disordered speech of the user and builds messages, which are converted into synthetic speech. System development was carried out employing user-centered design and development methods, which identified and refined key requirements for the device. A novel methodology for building small vocabulary, speaker-dependent automatic speech recognizers with reduced amounts of training data, was applied. Experiments showed that this method is successful in generating good recognition performance (mean accuracy 96%) on highly disordered speech, even when recognition perplexity is increased. The selected message-building technique traded off various factors including speed of message construction and range of available message outputs. The VIVOCA was evaluated in a field trial by individuals with moderate to severe dysarthria and confirmed that they can make use of the device to produce intelligible speech output from disordered speech input. The trial highlighted some issues which limit the performance and usability of the device when applied in real usage situations, with mean recognition accuracy of 67% in these circumstances. These limitations will be addressed in future work.

  7. A voice-input voice-output communication aid for people with severe speech impairment.

    PubMed

    Hawley, Mark S; Cunningham, Stuart P; Green, Phil D; Enderby, Pam; Palmer, Rebecca; Sehgal, Siddharth; O'Neill, Peter

    2013-01-01

    A new form of augmentative and alternative communication (AAC) device for people with severe speech impairment-the voice-input voice-output communication aid (VIVOCA)-is described. The VIVOCA recognizes the disordered speech of the user and builds messages, which are converted into synthetic speech. System development was carried out employing user-centered design and development methods, which identified and refined key requirements for the device. A novel methodology for building small vocabulary, speaker-dependent automatic speech recognizers with reduced amounts of training data, was applied. Experiments showed that this method is successful in generating good recognition performance (mean accuracy 96%) on highly disordered speech, even when recognition perplexity is increased. The selected message-building technique traded off various factors including speed of message construction and range of available message outputs. The VIVOCA was evaluated in a field trial by individuals with moderate to severe dysarthria and confirmed that they can make use of the device to produce intelligible speech output from disordered speech input. The trial highlighted some issues which limit the performance and usability of the device when applied in real usage situations, with mean recognition accuracy of 67% in these circumstances. These limitations will be addressed in future work. PMID:22875259

  8. Performance of a low data rate speech codec for land-mobile satellite communications

    NASA Technical Reports Server (NTRS)

    Gersho, Allen; Jedrey, Thomas C.

    1990-01-01

    In an effort to foster the development of new technologies for the emerging land mobile satellite communications services, JPL funded two development contracts in 1984: one to the Univ. of Calif., Santa Barbara and the other to the Georgia Inst. of Technology, to develop algorithms and real time hardware for near toll quality speech compression at 4800 bits per second. Both universities have developed and delivered speech codecs to JPL, and the UCSB codec was extensively tested by JPL in a variety of experimental setups. The basic UCSB speech codec algorithms and the test results of the various experiments performed with this codec are presented.

  9. A Robust Approach For Acoustic Noise Suppression In Speech Using ANFIS

    NASA Astrophysics Data System (ADS)

    Martinek, Radek; Kelnar, Michal; Vanus, Jan; Bilik, Petr; Zidek, Jan

    2015-11-01

    The authors of this article deals with the implementation of a combination of techniques of the fuzzy system and artificial intelligence in the application area of non-linear noise and interference suppression. This structure used is called an Adaptive Neuro Fuzzy Inference System (ANFIS). This system finds practical use mainly in audio telephone (mobile) communication in a noisy environment (transport, production halls, sports matches, etc). Experimental methods based on the two-input adaptive noise cancellation concept was clearly outlined. Within the experiments carried out, the authors created, based on the ANFIS structure, a comprehensive system for adaptive suppression of unwanted background interference that occurs in audio communication and degrades the audio signal. The system designed has been tested on real voice signals. This article presents the investigation and comparison amongst three distinct approaches to noise cancellation in speech; they are LMS (least mean squares) and RLS (recursive least squares) adaptive filtering and ANFIS. A careful review of literatures indicated the importance of non-linear adaptive algorithms over linear ones in noise cancellation. It was concluded that the ANFIS approach had the overall best performance as it efficiently cancelled noise even in highly noise-degraded speech. Results were drawn from the successful experimentation, subjective-based tests were used to analyse their comparative performance while objective tests were used to validate them. Implementation of algorithms was experimentally carried out in Matlab to justify the claims and determine their relative performances.

  10. Impact of internal waves on the coherent underwater acoustic communication

    NASA Astrophysics Data System (ADS)

    Wang, Jun; Yang, Xiaoxia; Wu, Lixin; Wang, Haibin; Lynch, James F.; Newhall, Arthur

    2012-11-01

    The 2001 Asian Sea International Experiment (ASIAEX2001) is a large-scale survey performed in the South China Sea. During May 2001 several minutes of M-sequence phase modulation signals were transmitted by a 400-Hz source repeatedly at intervals of half an hour, and received by an array 31 km away to conduct tomography of internal waves. The signal was treated as a binary-phase shift-keying (BPSK) communication signal with an information rate of 100 bps. The communication signals were demodulated by a decision-feedback equalizer. Since the intensity of the internal waves was not stable during the experiment period, data of two transmissions corresponding to a strong and a weak internal-wave activity were separately located and processed to investigate the impact of internal waves on the coherent underwater acoustic communication. The results show that internal waves cause a greatly fluctuating signal level and a rapidly varying multipath structure; consequently, these results show that the parameters of the equalizer need to be adjusted to mitigate the degradation of the communication performance.

  11. Implementing the Speech Showcase To Encourage and Display Diversity in a Speech Communication Department.

    ERIC Educational Resources Information Center

    Hall, Danford A.

    A "Speech Showcase," such as the one instituted over the past 5 years at Del Mar College (Texas), gives students enrolled in basic public speaking courses the opportunity to perform before a large audience and thereby gain experience and recognition. The faculty facilitator sets up the event logistically and advertises the event. Student speakers…

  12. Intergenerational communication: fundamental but under-exploited theory for speech and language therapy with older people.

    PubMed

    Armstrong, Linda; McKechnie, Karen

    2003-01-01

    There is a body of research literature already applied in speech and language therapy practice that is concerned with communication between children and adults and the adaptations adults make to facilitate the development of language in children. There is much less and more recent literature concerned with intergenerational communication involving older people and older people in institutional care. This has not yet impacted on speech and language therapy practice, especially in the area of training others. The aims of this paper are (1) to describe some of the main theoretical concepts associated with intergenerational communication, (2) to present the results from a study of the opinions about and experiences of intergenerational communication in which children, community- based older women and professional carers of older people were included and (3) to discuss the implications for speech and language therapy practice. A hypothesis for the study was that views on and attitudes towards communication and ageing would vary among the age groups. Some of the main concepts and models associated with intergenerational communication with older people are reviewed, including the communication predicament and enhancement models and the concept of patronizing communication. A qualitative study of three different age groups of (mainly) women was undertaken using a variety of methods of data elicitation (including written questionnaire and focus group discussion). Themes arising from the data were illuminated using content analysis. Participants' responses demonstrate some current generally positive views across the life-span on what it means to be old and the value of communication with older people. The implications for speech and language therapy practice are outlined, with the main emphasis on the potential use of intergenerational communication theory in developing a new focus for training other staff groups who care for older people and for measures of effectiveness of

  13. Acoustic source characteristics, across-formant integration, and speech intelligibility under competitive conditions.

    PubMed

    Roberts, Brian; Summers, Robert J; Bailey, Peter J

    2015-06-01

    An important aspect of speech perception is the ability to group or select formants using cues in the acoustic source characteristics--for example, fundamental frequency (F0) differences between formants promote their segregation. This study explored the role of more radical differences in source characteristics. Three-formant (F1+F2+F3) synthetic speech analogues were derived from natural sentences. In Experiment 1, F1+F3 were generated by passing a harmonic glottal source (F0 = 140 Hz) through second-order resonators (H1+H3); in Experiment 2, F1+F3 were tonal (sine-wave) analogues (T1+T3). F2 could take either form (H2 or T2). In some conditions, the target formants were presented alone, either monaurally or dichotically (left ear = F1+F3; right ear = F2). In others, they were accompanied by a competitor for F2 (F1+F2C+F3; F2), which listeners must reject to optimize recognition. Competitors (H2C or T2C) were created using the time-reversed frequency and amplitude contours of F2. Dichotic presentation of F2 and F2C ensured that the impact of the competitor arose primarily through informational masking. In the absence of F2C, the effect of a source mismatch between F1+F3 and F2 was relatively modest. When F2C was present, intelligibility was lowest when F2 was tonal and F2C was harmonic, irrespective of which type matched F1+F3. This finding suggests that source type and context, rather than similarity, govern the phonetic contribution of a formant. It is proposed that wideband harmonic analogues are more effective informational maskers than narrowband tonal analogues, and so become dominant in across-frequency integration of phonetic information when placed in competition. PMID:25751040

  14. Sine-wave and noise-vocoded sine-wave speech in a tone language: Acoustic details matter.

    PubMed

    Rosen, Stuart; Hui, Sze Ngar Catherine

    2015-12-01

    Sine-wave speech (SWS) is a highly simplified version of speech consisting only of frequency- and amplitude-modulated sinusoids representing the formants. That listeners can successfully understand SWS has led to claims that speech perception must be based on abstract properties of the stimuli far removed from their specific acoustic form. Here it is shown, in bilingual Cantonese/English listeners, that performance with Cantonese SWS is improved by noise vocoding, with no effect on English SWS utterances. This manipulation preserves the abstract informational structure in the signals but changes its surface form. The differential effects of noise vocoding likely arise from the fact that Cantonese is a tonal language and hence more reliant on fundamental frequency (F0) contours for its intelligibility. SWS does not preserve tonal information from the original speech but does have false tonal information signalled by the lowest frequency sinusoid. Noise vocoding SWS appears to minimise the tonal percept, which thus interferes less in the perception of Cantonese. It has no effect in English, which is minimally reliant on F0 variations for intelligibility. Therefore it is not only the informational structure of a sound that is important but also how its acoustic detail interacts with the phonological structure of a given language.

  15. Effects of a music therapy voice protocol on speech intelligibility, vocal acoustic measures, and mood of individuals with Parkinson's disease.

    PubMed

    Haneishi, E

    2001-01-01

    This study examined the effects of a Music Therapy Voice Protocol (MTVP) on speech intelligibility, vocal intensity, maximum vocal range, maximum duration of sustained vowel phonation, vocal fundamental frequency, vocal fundamental frequency variability, and mood of individuals with Parkinson's disease. Four female patients, who demonstrated voice and speech problems, served as their own controls and participated in baseline assessment (study pretest), a series of MTVP sessions involving vocal and singing exercises, and final evaluation (study posttest). In study pre and posttests, data for speech intelligibility and all acoustic variables were collected. Statistically significant increases were found in speech intelligibility, as rated by caregivers, and in vocal intensity from study pretest to posttest as the results of paired samples t-tests. In addition, before and after each MTVP session (session pre and posttests), self-rated mood scores and selected acoustic variables were collected. No significant differences were found in any of the variables from the session pretests to posttests, across the entire treatment period, or their interactions as the results of two-way ANOVAs with repeated measures. Although not significant, the mean of mood scores in session posttests (M = 8.69) was higher than that in session pretests (M = 7.93). PMID:11796078

  16. Sine-wave and noise-vocoded sine-wave speech in a tone language: Acoustic details matter.

    PubMed

    Rosen, Stuart; Hui, Sze Ngar Catherine

    2015-12-01

    Sine-wave speech (SWS) is a highly simplified version of speech consisting only of frequency- and amplitude-modulated sinusoids representing the formants. That listeners can successfully understand SWS has led to claims that speech perception must be based on abstract properties of the stimuli far removed from their specific acoustic form. Here it is shown, in bilingual Cantonese/English listeners, that performance with Cantonese SWS is improved by noise vocoding, with no effect on English SWS utterances. This manipulation preserves the abstract informational structure in the signals but changes its surface form. The differential effects of noise vocoding likely arise from the fact that Cantonese is a tonal language and hence more reliant on fundamental frequency (F0) contours for its intelligibility. SWS does not preserve tonal information from the original speech but does have false tonal information signalled by the lowest frequency sinusoid. Noise vocoding SWS appears to minimise the tonal percept, which thus interferes less in the perception of Cantonese. It has no effect in English, which is minimally reliant on F0 variations for intelligibility. Therefore it is not only the informational structure of a sound that is important but also how its acoustic detail interacts with the phonological structure of a given language. PMID:26723325

  17. Production and perception of clear speech

    NASA Astrophysics Data System (ADS)

    Bradlow, Ann R.

    2003-04-01

    When a talker believes that the listener is likely to have speech perception difficulties due to a hearing loss, background noise, or a different native language, she or he will typically adopt a clear speaking style. Previous research has established that, with a simple set of instructions to the talker, ``clear speech'' can be produced by most talkers under laboratory recording conditions. Furthermore, there is reliable evidence that adult listeners with either impaired or normal hearing typically find clear speech more intelligible than conversational speech. Since clear speech production involves listener-oriented articulatory adjustments, a careful examination of the acoustic-phonetic and perceptual consequences of the conversational-to-clear speech transformation can serve as an effective window into talker- and listener-related forces in speech communication. Furthermore, clear speech research has considerable potential for the development of speech enhancement techniques. After reviewing previous and current work on the acoustic properties of clear versus conversational speech, this talk will present recent data from a cross-linguistic study of vowel production in clear speech and a cross-population study of clear speech perception. Findings from these studies contribute to an evolving view of clear speech production and perception as reflecting both universal, auditory and language-specific, phonological contrast enhancement features.

  18. Evidence for acoustic communication among bottom foraging humpback whales.

    PubMed

    Parks, Susan E; Cusano, Dana A; Stimpert, Alison K; Weinrich, Mason T; Friedlaender, Ari S; Wiley, David N

    2014-12-16

    Humpback whales (Megaptera novaeangliae), a mysticete with a cosmopolitan distribution, demonstrate marked behavioural plasticity. Recent studies show evidence of social learning in the transmission of specific population level traits ranging from complex singing to stereotyped prey capturing behaviour. Humpback whales have been observed to employ group foraging techniques, however details on how individuals coordinate behaviour in these groups is challenging to obtain. This study investigates the role of a novel broadband patterned pulsed sound produced by humpback whales engaged in bottom-feeding behaviours, referred to here as a 'paired burst' sound. Data collected from 56 archival acoustic tag deployments were investigated to determine the functional significance of these signals. Paired burst sound production was associated exclusively with bottom feeding under low-light conditions, predominantly with evidence of associated conspecifics nearby suggesting that the sound likely serves either as a communicative signal to conspecifics, a signal to affect prey behaviour, or possibly both. This study provides additional evidence for individual variation and phenotypic plasticity of foraging behaviours in humpback whales and provides important evidence for the use of acoustic signals among foraging individuals in this species.

  19. Evidence for acoustic communication among bottom foraging humpback whales.

    PubMed

    Parks, Susan E; Cusano, Dana A; Stimpert, Alison K; Weinrich, Mason T; Friedlaender, Ari S; Wiley, David N

    2014-01-01

    Humpback whales (Megaptera novaeangliae), a mysticete with a cosmopolitan distribution, demonstrate marked behavioural plasticity. Recent studies show evidence of social learning in the transmission of specific population level traits ranging from complex singing to stereotyped prey capturing behaviour. Humpback whales have been observed to employ group foraging techniques, however details on how individuals coordinate behaviour in these groups is challenging to obtain. This study investigates the role of a novel broadband patterned pulsed sound produced by humpback whales engaged in bottom-feeding behaviours, referred to here as a 'paired burst' sound. Data collected from 56 archival acoustic tag deployments were investigated to determine the functional significance of these signals. Paired burst sound production was associated exclusively with bottom feeding under low-light conditions, predominantly with evidence of associated conspecifics nearby suggesting that the sound likely serves either as a communicative signal to conspecifics, a signal to affect prey behaviour, or possibly both. This study provides additional evidence for individual variation and phenotypic plasticity of foraging behaviours in humpback whales and provides important evidence for the use of acoustic signals among foraging individuals in this species. PMID:25512188

  20. Evidence for acoustic communication among bottom foraging humpback whales

    PubMed Central

    Parks, Susan E.; Cusano, Dana A.; Stimpert, Alison K.; Weinrich, Mason T.; Friedlaender, Ari S.; Wiley, David N.

    2014-01-01

    Humpback whales (Megaptera novaeangliae), a mysticete with a cosmopolitan distribution, demonstrate marked behavioural plasticity. Recent studies show evidence of social learning in the transmission of specific population level traits ranging from complex singing to stereotyped prey capturing behaviour. Humpback whales have been observed to employ group foraging techniques, however details on how individuals coordinate behaviour in these groups is challenging to obtain. This study investigates the role of a novel broadband patterned pulsed sound produced by humpback whales engaged in bottom-feeding behaviours, referred to here as a ‘paired burst' sound. Data collected from 56 archival acoustic tag deployments were investigated to determine the functional significance of these signals. Paired burst sound production was associated exclusively with bottom feeding under low-light conditions, predominantly with evidence of associated conspecifics nearby suggesting that the sound likely serves either as a communicative signal to conspecifics, a signal to affect prey behaviour, or possibly both. This study provides additional evidence for individual variation and phenotypic plasticity of foraging behaviours in humpback whales and provides important evidence for the use of acoustic signals among foraging individuals in this species. PMID:25512188

  1. Acoustics and sociolinguistics: Patterns of communication in hearing impairing classrooms

    NASA Astrophysics Data System (ADS)

    McKellin, William; Shahin, Kimary; Jamieson, Janet; Hodgson, Murray; Pichora-Fuller, Kathleen

    2005-04-01

    In elementary school classes, noise during student led activities is often taken as evidence of successful interaction and learning. In this complex social environment of elementary school classrooms, acquisition of complex language and social skills-the focus of activities in early education-is expected to take place in hearing-hostile environments. Communication and language processing in these contexts requires interactive strategies, discourse forms, and syntactic structures different from the educationally desired forms used in acoustically advantageous environments. Recordings were made of the interaction of groups of students in grades 1-3, 5, and 7 during collaborative group work in their regular classrooms. Each student wore microphones at the ear level and head-mounted video cameras. Each group as a whole was also audio- and videotaped and noise level readings were recorded. Analysis of the acoustical and phonological properties of language heard by each student has demonstrated that the language variety used in these noisy and reverberant settings is similar to that of individuals with hearing impairments. This paper reports similarities between the syntactic structures and pragmatic strategies used by hearing impaired children and normally hearing children in noisy contexts. [Work supported by Peter Wall Institute for Advanced Studies, University of British Columbia.

  2. Psychoacoustic Assessment of Speech Communication Systems. The Diagnostic Discrimination Test.

    ERIC Educational Resources Information Center

    Grether, Craig Blaine

    The present report traces the rationale, development and experimental evaluation of the Diagnostic Discrimination Test (DDT). The DDT is a three-choice test of consonant discriminability of the perceptual/acoustic dimensions of consonant phonemes within specific vowel contexts. The DDT was created and developed in an attempt to provide a…

  3. Multichannel decision feedback equalization in underwater acoustic communication

    NASA Astrophysics Data System (ADS)

    Yang, Xiaoxia; Wang, Haibin; Wang, Jun

    2012-11-01

    This paper studies an underwater acoustic communication system and its performance in shallow water. A channel encoder and a quadrature phase shift keying (QPSK) modulation scheme have been adopted at the transmitter, and a multichannel decision feedback equalizer (DFE) with a soft decision device has been used at the receiver. This system has been tested in a Yellow Sea experiment for transmitting an image at a range of 18.5km. A space diversity procedure is used to mitigate the performance degradation of the multichannel DFE caused by occasional impulsive noise in the receiver. In the case of two receivers, the equalizer exhibits good performance with bit error number of 319 out of 239946. After channel decoding, an image is obtained without any errant bits.

  4. Efficient use of bandwidth for underwater acoustic communication.

    PubMed

    Song, H C; Hodgkiss, W S

    2013-08-01

    In a recent shallow water experiment, acoustic communication transmissions were carried out over the 10 to 32 kHz band in ~100 m deep water over a 3 km range. A natural question is how best to utilize that bandwidth. In one multiband approach discussed previously, the band was divided into four smaller subbands that were processed independently using time reversal decision-feedback equalizers (TR-DFEs). This letter presents a complementary wideband approach using data from the same experiment achieving a data rate of up to 60 kbits/s with 32 quadrature amplitude modulation. These results suggest that a wideband approach can be beneficial in terms of spectral efficiency with modest computational complexity using a TR-DFE.

  5. Acoustic changes in the production of lexical stress during Lombard speech.

    PubMed

    Arciuli, Joanne; Simpson, Briony S; Vogel, Adam P; Ballard, Kirrie J

    2014-06-01

    The Lombard effect describes the phenomenon of individuals increasing their vocal intensity when speaking in the presence of background noise. Here, we conducted an investigation of the production of lexical stress during Lombard speech. Participants (N = 27) produced the same sentences in three conditions: one quiet condition and two noise conditions at 70 dB (white noise; multi-talker babble). Manual acoustic analyses (syllable duration, vowel intensity, and vowel fundamental frequency) were completed for repeated productions of two trisyllabic words with opposing patterns of lexical stress (weak-strong; strong-weak) in each of the three conditions. In total, 324 productions were analysed (12 utterances per participant). Results revealed that, rather than increasing vocal intensity equally across syllables, participants alter the degree of stress contrastivity when speaking in noise. This was especially evident in the production of strong-weak lexical stress where there was an increase in contrastivity across syllables in terms of intensity and fundamental frequency. This preliminary study paves the way for further research that is needed to establish these findings using a larger set of multisyllabic stimuli.

  6. Aerodynamic and acoustical measures of speech, operatic, and Broadway vocal styles in a professional female singer.

    PubMed

    Stone, R E; Cleveland, Thomas F; Sundberg, P Johan; Prokop, Jan

    2003-09-01

    Understanding how the voice is used in different styles of singing is commonly based on intuitive descriptions offered by performers who are proficient in only one style. Such descriptions are debatable, lack reproducibility, and lack scientifically derived explanations of the characteristics. We undertook acoustic and aerodynamic analyses of a female subject with professional experience in both operatic and Broadway styles of singing, who sang examples in these two styles. How representative the examples are of the respective styles was investigated by means of a listening test. Further, as a reference point, we compared the styles with her speech. Variation in styles associated with pitch and vocal loudness was investigated for various parameters: subglottal pressure, closed quotient, glottal leakage, H1-H2 difference (the level difference between the two lowest partials of the source spectrum), and glottal compliance (the ratio between the air volume displaced in a glottal pulse and the subglottal pressure). Formant frequencies, long-term-average spectrum, and vibrato characteristics were also studied. Characteristics of operatic style emerge as distinctly different from Broadway style, the latter being more similar to speaking. PMID:14513952

  7. Practical Considerations When Supporting Transitions for Pupils with Speech, Language and Communication Needs

    ERIC Educational Resources Information Center

    Perfitt, Ruth

    2013-01-01

    This article investigates the impact of transitions upon pupils aged 11-14 with speech, language and communication needs, including specific language impairment and autism. The aim is to identify stress factors, examine whether these affect any subgroups in particular and suggest practical strategies to support pupils through transitions. Stress…

  8. Preschool Children's Performance on Profiling Elements of Prosody in Speech-Communication (PEPS-C)

    ERIC Educational Resources Information Center

    Gibbon, Fiona E.; Smyth, Heather

    2013-01-01

    Profiling Elements of Prosody in Speech-Communication (PEPS-C) has not been used widely to assess prosodic abilities of preschool children. This study was therefore aimed at investigating typically developing 4-year-olds' performance on PEPS-C. PEPS-C was presented to 30 typically developing 4-year-olds recruited in southern Ireland. Children were…

  9. Report on a Pilot Project Which Combined Speech Communication and English Composition Instruction.

    ERIC Educational Resources Information Center

    Friday, Robert A.; Beranek, Bernard F.

    In reorganizing its core curriculum, Duquesne University (Pennsylvania) created a course that combined the required freshman composition and speech communication courses. A two-term pilot course offered eight hours of credit, meeting for one hour three times a week and for a two hour lab once a week. Oral readings were assigned in order to help…

  10. Altered Gesture and Speech Production in ASD Detract from In-Person Communicative Quality

    ERIC Educational Resources Information Center

    Morett, Laura M.; O'Hearn, Kirsten; Luna, Beatriz; Ghuman, Avniel Singh

    2016-01-01

    This study disentangled the influences of language and social processing on communication in autism spectrum disorder (ASD) by examining whether gesture and speech production differs as a function of social context. The results indicate that, unlike other adolescents, adolescents with ASD did not increase their coherency and engagement in the…

  11. The Primary Aim of Speech Communication Education: Some Potential Concerns for Members of the Discipline.

    ERIC Educational Resources Information Center

    Hudson, David D.

    Following a brief discussion of the works of the philosopher Jacques Maritain, who viewed the educational process as the art of helping individuals to realize the nature of that which is inherent in them, this paper analyzes the implications of Maritain's views for the discipline of speech communication. The paper suggests that education in speech…

  12. Meeting the Needs of Children and Young People with Speech, Language and Communication Difficulties

    ERIC Educational Resources Information Center

    Lindsay, Geoff; Dockrell, Julie; Desforges, Martin; Law, James; Peacey, Nick

    2010-01-01

    Background: The UK government set up a review of provision for children and young people with the full range of speech, language and communication needs led by a Member of Parliament, John Bercow. A research study was commissioned to provide empirical evidence to inform the Bercow Review. Aims: To examine the efficiency and effectiveness of…

  13. Creative Speech Communication for South Dakota. K-12 Curriculum Resource Book.

    ERIC Educational Resources Information Center

    South Dakota Speech Association.

    Intended for regular elementary and secondary teachers, the South Dakota curriculum guide presents motivating activities to stimulate speech development of under-communicating, gifted, experientially different, or average students. Suggested situations are reported to be coordinated with South Dakota literature, the best of children's literature,…

  14. Communicating Epistemic Stance: How Speech and Gesture Patterns Reflect Epistemicity and Evidentiality

    ERIC Educational Resources Information Center

    Roseano, Paolo; González, Montserrat; Borràs-Comes, Joan; Prieto, Pilar

    2016-01-01

    This study investigates how epistemic stance is encoded and perceived in face-to-face communication when language is regarded as comprised by speech and gesture. Two studies were conducted with this goal in mind. The first study consisted of a production task in which participants performed opinion reports. Results showed that speakers communicate…

  15. DIAGNOSIS AND APPRAISAL OF COMMUNICATION DISORDERS. PRENTICE-HALL FOUNDATIONS OF SPEECH PATHOLOGY SERIES.

    ERIC Educational Resources Information Center

    DARLEY, FREDERIC L.

    THIS TEXT GIVES THE STUDENT AN OUTLINE OF THE BASIC PRINCIPLES OF SCIENTIFIC METHODOLOGY WHICH UNDERLIE EVALUATIVE WORK IN SPEECH DISORDERS. RATIONALE AND ASSESSMENT TECHNIQUES ARE GIVEN FOR EXAMINATION OF THE BASIC COMMUNICATION PROCESSES OF SYMBOLIZATION, RESPIRATION, PHONATION, ARTICULATION-RESONANCE, PROSODY, ASSOCIATED SENSORY AND PERCEPTUAL…

  16. Communication and Behavior: Collaboration between Speech-Language Pathologists and Behavioral Psychologists.

    ERIC Educational Resources Information Center

    Ylvisaker, Mark; Feeney, Timothy J.

    1994-01-01

    This article presents an intervention framework in which speech-language pathologists and behavioral psychologists view the challenging behaviors of individuals with traumatic brain injury from a communication perspective and address these problems in a way that is positive, integrated, functional, collaborative, and environmental. (Author/JDD)

  17. ARAMCO Education: Teaching Speech Communication to a Sub-Culture in Saudi Arabia.

    ERIC Educational Resources Information Center

    Dick, Robert C.

    Based on experiences gained by an educator from Indiana University who taught a speech communication course in Saudi Arabia, this paper details the adaptations the educator had to make in order to teach Arabian American Oil Company (ARAMCO) employees and their spouses in the politically difficult period of 1981-82. Following a brief background…

  18. Speech Intelligibility and Marital Communication in Amyotrophic Lateral Sclerosis: An Exploratory Study

    ERIC Educational Resources Information Center

    Joubert, Karin; Bornman, Juan; Alant, Erna

    2011-01-01

    Amyotrophic lateral sclerosis (ALS), a rapidly progressive neuromuscular disease, has a devastating impact not only on individuals diagnosed with ALS but also their spouses. Speech intelligibility, often compromised as a result of dysarthria, affects the couple's ability to maintain effective, intimate communication. The purpose of this…

  19. Age-related Changes in Acoustic Modifications of Mandarin Maternal Speech to Preverbal Infants and Five-Year-Old Children: A Longitudinal Study

    PubMed Central

    Liu, Huei-Mei; Tsao, Feng-Ming; Kuhl, Patricia K.

    2010-01-01

    Acoustic-phonetic exaggeration of infant-directed speech (IDS) is well documented, but few studies address whether these features are modified with a child's age. Mandarin-speaking mothers were recorded while addressing an adult and their child at two ages (7-12 months and 5 years) to examine the acoustic-phonetic differences between IDS and child–directed speech (CDS). CDS exhibits an exaggeration pattern resembling that of IDS—expanded vowel space, longer vowels, higher pitch, and greater lexical tone differences—when compared to ADS. Longitudinal analysis demonstrated that the extent of acoustic exaggeration is significantly smaller in CDS than in IDS. Age-related changes in maternal speech provide some support for the hypothesis that mothers adjust their speech directed toward children as a function of the child's language ability. PMID:19232142

  20. Adopting public health approaches to communication disability: challenges for the education of speech-language pathologists.

    PubMed

    Wylie, Karen; McAllister, Lindy; Davidson, Bronwyn; Marshall, Julie; Law, James

    2014-01-01

    Public health approaches to communication disability challenge the profession of speech-language pathology (SLP) to reconsider both frames of reference for practice and models of education. This paper reviews the impetus for public health approaches to communication disability and considers how public health is, and could be, incorporated into SLP education, both now and in the future. The paper describes tensions between clinical services, which have become increasingly specialized, and public health approaches that offer a broader view of communication disability and communication disability prevention. It presents a discussion of these tensions and asserts that public health approaches to communication are themselves a specialist field, requiring specific knowledge and skills. The authors suggest the use of the term 'communication disability public health' to refer to this type of work and offer a preliminary definition in order to advance discussion. Examples from three countries are provided of how some SLP degree programmes are integrating public health into the SLP curriculum. Alternative models of training for communication disability public health that may be relevant in the future in different contexts and countries are presented, prompting the SLP profession to consider whether communication disability public health is a field of practice for speech-language pathologists or whether it has broader workforce implications. The paper concludes with some suggestions for the future which may advance thinking, research and practice in communication disability public health.

  1. Low complexity adaptive equalizers for underwater acoustic communications

    NASA Astrophysics Data System (ADS)

    Soflaei, Masoumeh; Azmi, Paeiz

    2014-08-01

    Interference signals due to scattering from surface and reflecting from bottom is one of the most important problems of reliable communications in shallow water channels. To solve this problem, one of the best suggested ways is to use adaptive equalizers. Convergence rate and misadjustment error in adaptive algorithms play important roles in adaptive equalizer performance. In this paper, affine projection algorithm (APA), selective regressor APA(SR-APA), family of selective partial update (SPU) algorithms, family of set-membership (SM) algorithms and selective partial update selective regressor APA (SPU-SR-APA) are compared with conventional algorithms such as the least mean square (LMS) in underwater acoustic communications. We apply experimental data from the Strait of Hormuz for demonstrating the efficiency of the proposed methods over shallow water channel. We observe that the values of the steady-state mean square error (MSE) of SR-APA, SPU-APA, SPU-normalized least mean square (SPU-NLMS), SPU-SR-APA, SM-APA and SM-NLMS algorithms decrease in comparison with the LMS algorithm. Also these algorithms have better convergence rates than LMS type algorithm.

  2. Language-Specific Developmental Differences in Speech Production: A Cross-Language Acoustic Study

    ERIC Educational Resources Information Center

    Li, Fangfang

    2012-01-01

    Speech productions of 40 English- and 40 Japanese-speaking children (aged 2-5) were examined and compared with the speech produced by 20 adult speakers (10 speakers per language). Participants were recorded while repeating words that began with "s" and "sh" sounds. Clear language-specific patterns in adults' speech were found, with English…

  3. A picture communicator for symbol users and/or speech-impaired people.

    PubMed

    Blenkhorn, P

    1992-01-01

    There are several approaches to producing communication aids for people with disabilities. The system described here takes the approach of utilizing as much mainstream hardware as possible, and adapting it with some modular software tools which have been designed to facilitate the building of customized symbol communication systems. The target audience are clients who are symbol users and/or have a speech impairment. The system provides several levels of screens, each of which can contain grids of scalable icons. A number of input methods are provided including keyboard, switch, mouse and touch-screen. Digitized and/or text to speech synthesis can be used for reinforcement of selections and for communication. The structure of the system is discussed and initial feedback from the first field trials is presented. PMID:1300378

  4. Altered Gesture and Speech Production in ASD Detract from In-Person Communicative Quality.

    PubMed

    Morett, Laura M; O'Hearn, Kirsten; Luna, Beatriz; Ghuman, Avniel Singh

    2016-03-01

    This study disentangled the influences of language and social processing on communication in autism spectrum disorder (ASD) by examining whether gesture and speech production differs as a function of social context. The results indicate that, unlike other adolescents, adolescents with ASD did not increase their coherency and engagement in the presence of a visible listener, and that greater coherency and engagement were related to lesser social and communicative impairments. Additionally, the results indicated that adolescents with ASD produced sparser speech and fewer gestures conveying supplementary information, and that both of these effects increased in the presence of a visible listener. Together, these findings suggest that interpersonal communication deficits in ASD are driven more strongly by social processing than language processing.

  5. Acoustic Communication at the Water's Edge: Evolutionary Insights from a Mudskipper

    PubMed Central

    Polgar, Gianluca; Malavasi, Stefano; Cipolato, Giacomo; Georgalas, Vyron; Clack, Jennifer A.; Torricelli, Patrizia

    2011-01-01

    Coupled behavioural observations and acoustical recordings of aggressive dyadic contests showed that the mudskipper Periophthalmodon septemradiatus communicates acoustically while out of water. An analysis of intraspecific variability showed that specific acoustic components may act as tags for individual recognition, further supporting the sounds' communicative value. A correlative analysis amongst acoustical properties and video-acoustical recordings in slow-motion supported first hypotheses on the emission mechanism. Acoustic transmission through the wet exposed substrate was also discussed. These observations were used to support an “exaptation hypothesis”, i.e. the maintenance of key adaptations during the first stages of water-to-land vertebrate eco-evolutionary transitions (based on eco-evolutionary and palaeontological considerations), through a comparative bioacoustic analysis of aquatic and semiterrestrial gobiid taxa. In fact, a remarkable similarity was found between mudskipper vocalisations and those emitted by gobioids and other soniferous benthonic fishes. PMID:21738663

  6. Acoustic and perceptual evaluation of category goodness of /t/ and /k/ in typical and misarticulated children's speech.

    PubMed

    Strömbergsson, Sofia; Salvi, Giampiero; House, David

    2015-06-01

    This investigation explores perceptual and acoustic characteristics of children's successful and unsuccessful productions of /t/ and /k/, with a specific aim of exploring perceptual sensitivity to phonetic detail, and the extent to which this sensitivity is reflected in the acoustic domain. Recordings were collected from 4- to 8-year-old children with a speech sound disorder (SSD) who misarticulated one of the target plosives, and compared to productions recorded from peers with typical speech development (TD). Perceptual responses were registered with regards to a visual-analog scale, ranging from "clear [t]" to "clear [k]." Statistical models of prototypical productions were built, based on spectral moments and discrete cosine transform features, and used in the scoring of SSD productions. In the perceptual evaluation, "clear substitutions" were rated as less prototypical than correct productions. Moreover, target-appropriate productions of /t/ and /k/ produced by children with SSD were rated as less prototypical than those produced by TD peers. The acoustical modeling could to a large extent discriminate between the gross categories /t/ and /k/, and scored the SSD utterances on a continuous scale that was largely consistent with the category of production. However, none of the methods exhibited the same sensitivity to phonetic detail as the human listeners.

  7. Resonant acoustic transducer and driver system for a well drilling string communication system

    DOEpatents

    Chanson, Gary J.; Nicolson, Alexander M.

    1981-01-01

    The acoustic data communication system includes an acoustic transmitter and receiver wherein low frequency acoustic waves, propagating in relatively loss free manner in well drilling string piping, are efficiently coupled to the drill string and propagate at levels competitive with the levels of noise generated by drilling machinery also present in the drill string. The transmitting transducer incorporates a mass-spring piezoelectric transmitter and amplifier combination that permits self-oscillating resonant operation in the desired low frequency range.

  8. Applications for Subvocal Speech

    NASA Technical Reports Server (NTRS)

    Jorgensen, Charles; Betts, Bradley

    2007-01-01

    A research and development effort now underway is directed toward the use of subvocal speech for communication in settings in which (1) acoustic noise could interfere excessively with ordinary vocal communication and/or (2) acoustic silence or secrecy of communication is required. By "subvocal speech" is meant sub-audible electromyographic (EMG) signals, associated with speech, that are acquired from the surface of the larynx and lingual areas of the throat. Topics addressed in this effort include recognition of the sub-vocal EMG signals that represent specific original words or phrases; transformation (including encoding and/or enciphering) of the signals into forms that are less vulnerable to distortion, degradation, and/or interception; and reconstruction of the original words or phrases at the receiving end of a communication link. Potential applications include ordinary verbal communications among hazardous- material-cleanup workers in protective suits, workers in noisy environments, divers, and firefighters, and secret communications among law-enforcement officers and military personnel in combat and other confrontational situations.

  9. Dual Perspectives on Theory in Clinical Practice: Practice Makes Perfect: The Incompatibility of Practicing Speech and Meaningful Communication.

    ERIC Educational Resources Information Center

    Kamhi, Alan G.

    2000-01-01

    This article uses a case study to suggest that some children view speech-language therapy as a separate situation for learning practicing new sounds and language forms whereas the purpose of talking outside of therapy is meaningful communication. Clinical implications of this potential incompatibility between practicing speech and communicating…

  10. Comparative Efficacy of the Picture Exchange Communication System (PECS) versus a Speech-Generating Device: Effects on Requesting Skills

    ERIC Educational Resources Information Center

    Boesch, Miriam C.; Wendt, Oliver; Subramanian, Anu; Hsu, Ning

    2013-01-01

    An experimental, single-subject research study investigated the comparative efficacy of the Picture Exchange Communication System (PECS) versus a speech-generating device (SGD) in developing requesting skills for three elementary-age children with severe autism and little to no functional speech. Results demonstrated increases in requesting…

  11. An Exploration of the Potential of Automatic Speech Recognition to Assist and Enable Receptive Communication in Higher Education

    ERIC Educational Resources Information Center

    Wald, Mike

    2006-01-01

    The potential use of Automatic Speech Recognition to assist receptive communication is explored. The opportunities and challenges that this technology presents students and staff to provide captioning of speech online or in classrooms for deaf or hard of hearing students and assist blind, visually impaired or dyslexic learners to read and search…

  12. Identifying the Challenges and Opportunities to Meet the Needs of Children with Speech, Language and Communication Difficulties

    ERIC Educational Resources Information Center

    Dockrell, Julie E.; Howell, Peter

    2015-01-01

    The views of experienced educational practitioners were examined with respect to the terminology used to describe children with speech, language and communication needs (SLCN), associated problems and the impact of speech and language difficulties in the classroom. Results showed that education staff continue to experience challenges with the…

  13. A Pilot Study on the Efficacy of Melodic Based Communication Therapy for Eliciting Speech in Nonverbal Children with Autism

    ERIC Educational Resources Information Center

    Sandiford, Givona A.; Mainess, Karen J.; Daher, Noha S.

    2013-01-01

    The purpose of this study was to compare the efficacy of Melodic Based Communication Therapy (MBCT) to traditional speech and language therapy for eliciting speech in nonverbal children with autism. Participants were 12 nonverbal children with autism ages 5 through 7 randomly assigned to either treatment group. Both groups made significant…

  14. Profiling Early Socio-Communicative Development in Five Young Girls with the Preserved Speech Variant of Rett Syndrome

    ERIC Educational Resources Information Center

    Marschik, Peter B.; Kaufmann, Walter E.; Einspieler, Christa; Bartl-Pokorny, Katrin D.; Wolin, Thomas; Pini, Giorgio; Budimirovic, Dejan B.; Zappella, Michele; Sigafoos, Jeff

    2012-01-01

    Rett syndrome (RTT) is a developmental disorder characterized by regression of purposeful hand skills and spoken language, although some affected children retain some ability to speech. We assessed the communicative abilities of five young girls, who were later diagnosed with the preserved speech variant of RTT, during the pre-regression period…

  15. Communicatively Developing Technology of Speech Activation of Tatar Schools' Teachers

    ERIC Educational Resources Information Center

    Khairova, Irina V.; Egamberdieva, Nodira M.

    2016-01-01

    The timeliness of the problem under study is that in conditions of the renewal of the content and system of education the interest to the communicative and development potentials of education in the system of course retraining of teachers who work in national (Tatar) schools is growing. However, actual academic complexes for teachers do not always…

  16. Simultaneous Communication: An Alternative to Speech Training with Autistic Children

    ERIC Educational Resources Information Center

    Konstantareas, M. Mary

    1977-01-01

    A similtaneous verbal and gestural approach to communication was used with 11 autistic children (5 to 10 years old). Available from: Journal of Practical Approaches to Developmental Handicap, 3304-33rd Street N.W., Calgary, Alberta, Canada T2L 2A6. (SBH)

  17. Convergent Aspects of Acoustic Communication in Darters, Sculpins, and Gobies.

    PubMed

    Zeyl, Jeffrey N; Malavasi, Stefano; Holt, Daniel E; Noel, Patricia; Lugli, Marco; Johnston, Carol E

    2016-01-01

    Darters (Perciformes, Percidae), sculpins (Perciformes, Cottidae), and gobioids (Gobiiformes, Gobioidei) exhibit convergent life history traits, including a benthic lifestyle and a cavity nesting spawning mode. Soniferous species within these taxa produce pulsed and/or tonal sounds with peak frequencies below 200 Hz (with some exceptions), primarily in agonistic and/or reproductive contexts. The reduced or absent swim bladders found in these taxa limit or prevent both hearing enhancement via pressure sensitivity and acoustic amplification of the contracting sonic muscles, which are associated with the skull and pectoral girdle. While such anatomies constrain communication to low frequency channels, optimization of the S/N (signal-to-noise) ratio in low frequency channels is evident for some gobies, as measured by habitat soundscape frequency windows, nest cavity sound amplification, and audiograms. Similar S/N considerations are applicable to many darter and sculpin systems. This chapter reviews the currently documented diversity of sound production in darters, sculpins, and gobioids within a phylogenetic context, examines the efficacy of signal transmission from senders to receivers (sound production mechanisms, audiograms, and masking challenges), and evaluates the potential functional significance of sound attributes in relation to territorial and reproductive behaviours. PMID:26515312

  18. An international perspective: supporting adolescents with speech, language, and communication needs in the United Kingdom.

    PubMed

    Joffe, Victoria

    2015-02-01

    This article provides an overview of the education system in the United Kingdom, with a particular focus on the secondary school context and supporting older children and young people with speech, language, and communication needs (SLCNs). Despite the pervasive nature of speech, language, and communication difficulties and their long-term impact on academic performance, mental health, and well-being, evidence suggests that there is limited support to older children and young people with SLCNs in the United Kingdom, relative to what is available in the early years. Focus in secondary schools is predominantly on literacy, with little attention to supporting oral language. The article provides a synopsis of the working practices of pediatric speech and language therapists working with adolescents in the United Kingdom and the type and level of speech and language therapy support provided for older children and young people with SLCNs in secondary and further education. Implications for the nature and type of specialist support to adolescents and adults with SLCNs are discussed.

  19. An international perspective: supporting adolescents with speech, language, and communication needs in the United Kingdom.

    PubMed

    Joffe, Victoria

    2015-02-01

    This article provides an overview of the education system in the United Kingdom, with a particular focus on the secondary school context and supporting older children and young people with speech, language, and communication needs (SLCNs). Despite the pervasive nature of speech, language, and communication difficulties and their long-term impact on academic performance, mental health, and well-being, evidence suggests that there is limited support to older children and young people with SLCNs in the United Kingdom, relative to what is available in the early years. Focus in secondary schools is predominantly on literacy, with little attention to supporting oral language. The article provides a synopsis of the working practices of pediatric speech and language therapists working with adolescents in the United Kingdom and the type and level of speech and language therapy support provided for older children and young people with SLCNs in secondary and further education. Implications for the nature and type of specialist support to adolescents and adults with SLCNs are discussed. PMID:25633146

  20. Acoustics

    NASA Technical Reports Server (NTRS)

    Goodman, Jerry R.; Grosveld, Ferdinand

    2007-01-01

    The acoustics environment in space operations is important to maintain at manageable levels so that the crewperson can remain safe, functional, effective, and reasonably comfortable. High acoustic levels can produce temporary or permanent hearing loss, or cause other physiological symptoms such as auditory pain, headaches, discomfort, strain in the vocal cords, or fatigue. Noise is defined as undesirable sound. Excessive noise may result in psychological effects such as irritability, inability to concentrate, decrease in productivity, annoyance, errors in judgment, and distraction. A noisy environment can also result in the inability to sleep, or sleep well. Elevated noise levels can affect the ability to communicate, understand what is being said, hear what is going on in the environment, degrade crew performance and operations, and create habitability concerns. Superfluous noise emissions can also create the inability to hear alarms or other important auditory cues such as an equipment malfunctioning. Recent space flight experience, evaluations of the requirements in crew habitable areas, and lessons learned (Goodman 2003; Allen and Goodman 2003; Pilkinton 2003; Grosveld et al. 2003) show the importance of maintaining an acceptable acoustics environment. This is best accomplished by having a high-quality set of limits/requirements early in the program, the "designing in" of acoustics in the development of hardware and systems, and by monitoring, testing and verifying the levels to ensure that they are acceptable.

  1. Fifty years of progress in acoustic phonetics

    NASA Astrophysics Data System (ADS)

    Stevens, Kenneth N.

    2004-10-01

    Three events that occurred 50 or 60 years ago shaped the study of acoustic phonetics, and in the following few decades these events influenced research and applications in speech disorders, speech development, speech synthesis, speech recognition, and other subareas in speech communication. These events were: (1) the source-filter theory of speech production (Chiba and Kajiyama; Fant); (2) the development of the sound spectrograph and its interpretation (Potter, Kopp, and Green; Joos); and (3) the birth of research that related distinctive features to acoustic patterns (Jakobson, Fant, and Halle). Following these events there has been systematic exploration of the articulatory, acoustic, and perceptual bases of phonological categories, and some quantification of the sources of variability in the transformation of this phonological representation of speech into its acoustic manifestations. This effort has been enhanced by studies of how children acquire language in spite of this variability and by research on speech disorders. Gaps in our knowledge of this inherent variability in speech have limited the directions of applications such as synthesis and recognition of speech, and have led to the implementation of data-driven techniques rather than theoretical principles. Some examples of advances in our knowledge, and limitations of this knowledge, are reviewed.

  2. The Effects of Home-Based Literacy Activities on the Communication of Students with Severe Speech and Motor Impairments

    ERIC Educational Resources Information Center

    Cox, Amy Swartz; Clark, Denise M.; Skoning, Stacey N.; Wegner, Theresa M.; Muwana, Florence C.

    2015-01-01

    This study examined the effects of using sensory, augmentative, and alternative communication (AAC), and supportive communication strategies on the rate and type of communication used by three students with severe speech and motor impairments (SSMI). Using a multiple baseline across behaviour design with sensory and AAC intervention phases,…

  3. Acoustic communication in Okanagana rimosa (Say) (Homoptera: Cicadidae).

    PubMed

    Stölting, Heiko; Moore, Thomas E; Lakes-Harlan, Reinhard

    2004-01-01

    The cicada Okanagana rimosa (Say) has an acoustic communication system with three types of loud timbal sounds: (i) A calling song lasting several seconds to about 1 min which consists of a sequence of chirps at a repetition rate of 83 chirps per second. Each chirp of about 6 ms duration contains 4-5 pulses. The sound level of the calling song is 87-90 dB SPL at a distance of 15 cm. (ii) An amplitude modulated courtship song with increasing amplitude and repetition rate of chirps and pulses. (iii) A protest squawk with irregular chirp and pulse structure. The spectra of all three types are similar and show main energy peaks at 8-10 kHz. Only males sing, and calling song production is influenced by the songs of other males, resulting in an almost continuous sound in dense populations. In such populations, the calling songs overlap and the temporal structure of individual songs is obscured within the habitat. The calling song of the broadly sympatric, closely related species O. canadensis (Provander) is similar in frequency content, but distinct in the temporal pattern (24 chirps per second, 24 ms chirp duration, eight pulses per chirp) which is likely important for species separation in sympatric populations. The hearing threshold of the auditory nerve is similar for females and males of O. rimosa and most sensitive at 4-5 kHz. Experiments in the field show that female phonotaxis of O. rimosa depends on parameters of the calling song. Most females are attracted to calling song models with a 9 kHz carrier frequency (peak frequency of the calling song), but not to models with a 5 kHz carrier frequency (minimum hearing threshold). Phonotaxis depends on temporal parameters of the conspecific song, especially chirp repetition rate. Calling song production is influenced by environmental factors, and likelihood to sing increases with temperature and brightness of the sky. Correspondingly, females perform phonotaxis most often during sunny conditions with temperatures above

  4. Speech Communication and Communication Processes: Abstracts of Doctoral Dissertations Published in "Dissertation Abstracts International," January through March 1978 (Vol. 38 Nos. 7 through 9).

    ERIC Educational Resources Information Center

    ERIC Clearinghouse on Reading and Communication Skills, Urbana, IL.

    This collection of abstracts is part of a continuing series providing information on recent doctoral dissertations. The 18 titles deal with a variety of topics, including the following: the relationship of speech communication variables to perceived communication competence; interpersonal communication in a simulation game of intercultural…

  5. Effectiveness of the Picture Exchange Communication System (PECS) on Communication and Speech for Children with Autism Spectrum Disorders: A Meta-Analysis

    ERIC Educational Resources Information Center

    Flippin, Michelle; Reszka, Stephanie; Watson, Linda R.

    2010-01-01

    Purpose: The Picture Exchange Communication System (PECS) is a popular communication-training program for young children with autism spectrum disorders (ASD). This meta-analysis reviews the current empirical evidence for PECS in affecting communication and speech outcomes for children with ASD. Method: A systematic review of the literature on PECS…

  6. Designing acoustics for linguistically diverse classrooms: Effects of background noise, reverberation and talker foreign accent on speech comprehension by native and non-native English-speaking listeners

    NASA Astrophysics Data System (ADS)

    Peng, Zhao Ellen

    The current classroom acoustics standard (ANSI S12.60-2010) recommends core learning spaces not to exceed background noise level (BNL) of 35 dBA and reverberation time (RT) of 0.6 second, based on speech intelligibility performance mainly by the native English-speaking population. Existing literature has not correlated these recommended values well with student learning outcomes. With a growing population of non-native English speakers in American classrooms, the special needs for perceiving degraded speech among non-native listeners, either due to realistic room acoustics or talker foreign accent, have not been addressed in the current standard. This research seeks to investigate the effects of BNL and RT on the comprehension of English speech from native English and native Mandarin Chinese talkers as perceived by native and non-native English listeners, and to provide acoustic design guidelines to supplement the existing standard. This dissertation presents two studies on the effects of RT and BNL on more realistic classroom learning experiences. How do native and non-native English-speaking listeners perform on speech comprehension tasks under adverse acoustic conditions, if the English speech is produced by talkers of native English (Study 1) versus native Mandarin Chinese (Study 2)? Speech comprehension materials were played back in a listening chamber to individual listeners: native and non-native English-speaking in Study 1; native English, native Mandarin Chinese, and other non-native English-speaking in Study 2. Each listener was screened for baseline English proficiency level, and completed dual tasks simultaneously involving speech comprehension and adaptive dot-tracing under 15 acoustic conditions, comprised of three BNL conditions (RC-30, 40, and 50) and five RT scenarios (0.4 to 1.2 seconds). The results show that BNL and RT negatively affect both objective performance and subjective perception of speech comprehension, more severely for non

  7. Acoustic Analysis of Clear Versus Conversational Speech in Individuals with Parkinson Disease

    ERIC Educational Resources Information Center

    Goberman, A.M.; Elmer, L.W.

    2005-01-01

    A number of studies have been devoted to the examination of clear versus conversational speech in non-impaired speakers. The purpose of these previous studies has been primarily to help increase speech intelligibility for the benefit of hearing-impaired listeners. The goal of the present study was to examine differences between conversational and…

  8. A Computational Model of Word Segmentation from Continuous Speech Using Transitional Probabilities of Atomic Acoustic Events

    ERIC Educational Resources Information Center

    Rasanen, Okko

    2011-01-01

    Word segmentation from continuous speech is a difficult task that is faced by human infants when they start to learn their native language. Several studies indicate that infants might use several different cues to solve this problem, including intonation, linguistic stress, and transitional probabilities between subsequent speech sounds. In this…

  9. Acoustic Analysis of the Speech of Children with Cochlear Implants: A Longitudinal Study

    ERIC Educational Resources Information Center

    Liker, Marko; Mildner, Vesna; Sindija, Branka

    2007-01-01

    The aim of the study was to analyse the speech of the children with cochlear implants, and compare it with the speech of hearing controls. We focused on three categories of Croatian sounds: vowels (F1 and F2 frequencies), fricatives (noise frequencies of /s/ and /[esh]/ ), and affricates (total duration and the pattern of stop-fricative components…

  10. Profiling early socio-communicative development in five young girls with the preserved speech variant of Rett syndrome.

    PubMed

    Marschik, Peter B; Kaufmann, Walter E; Einspieler, Christa; Bartl-Pokorny, Katrin D; Wolin, Thomas; Pini, Giorgio; Budimirovic, Dejan B; Zappella, Michele; Sigafoos, Jeff

    2012-01-01

    Rett syndrome (RTT) is a developmental disorder characterized by regression of purposeful hand skills and spoken language, although some affected children retain some ability to speech. We assessed the communicative abilities of five young girls, who were later diagnosed with the preserved speech variant of RTT, during the pre-regression period (aged 12-24 months). Videotapes, obtained by parents during routine family situations and celebrations, were analyzed to identify communicative forms and functions used by these toddlers. Non-verbal communicative forms dominated over verbal-communicative forms for six of the eight identified communication functions. Although the girls used various non-verbal forms to make requests, for example, none of the individuals were observed to make choices or request information. Early peculiarities in the speech-language domain during the first year of life became more prominent and evident during the second year of life as general differences between typical development and atypical development become more obvious in RTT. These findings highlight the importance of assessing socio-communicative forms and functions at early age in children with RTT. The results suggest that speech-language functions did not appear to play a major role in the children's communicative attempts. We conclude that, even among children with the preserved speech variant, socio-communicative deficits are present before regression and persist after this period.

  11. Study of Doppler Shift Correction for Underwater Acoustic Communication Using Orthogonal Signal Division Multiplexing

    NASA Astrophysics Data System (ADS)

    Ebihara, Tadashi; Mizutani, Keiichi

    2011-07-01

    In this study, we apply Doppler shift correction schemes for underwater acoustic (UWA) communication with orthogonal signal division multiplexing (OSDM) to achieve stable communication in underwater acoustic channels. Three Doppler correction schemes, which exploit the guard interval, are applied to UWA communication with OSDM and evaluated in simulations. Through a simulation in which only the Doppler effect is considered, we confirmed that by adapting schemes to UWA communication with OSDM, we can correct large Doppler shifts, which addresses the usual speed of vehicles and ships. Moreover, by considering both the Doppler effect and channel reverberation, we propose the best possible combination of Doppler correction schemes for UWA communication with OSDM. The results suggest that UWA communication with OSDM may lead to high-quality communication by considering channel reverberation and large Doppler shifts.

  12. Communication Patterns of Youth with Mental Retardation with and without Their Speech-Output Communication Devices.

    ERIC Educational Resources Information Center

    Romski, Mary Ann; Sevcik, Rose A.; Adamson, Lauren B.

    1999-01-01

    The communication skills of 13 youth with moderate or severe mental retardation were compared while they communicated with a "standard partner" with and without access to their augmented communication devices. With the devices, subjects communicated more information, more clearly, and more specifically than without the devices. (DB)

  13. Acoustic correlates of vowel intelligibility in clear and conversational speech for young normal-hearing and elderly hearing-impaired listeners.

    PubMed

    Ferguson, Sarah Hargus; Quené, Hugo

    2014-06-01

    The present investigation carried out acoustic analyses of vowels in clear and conversational speech produced by 41 talkers. Mixed-effects models were then deployed to examine relationships among acoustic and perceptual data for these vowels. Acoustic data include vowel duration, steady-state formant frequencies, and two measures of dynamic formant movement. Perceptual data consist of vowel intelligibility in noise for young normal-hearing and elderly hearing-impaired listeners, as reported by Ferguson in 2004 and 2012 [J. Acoust. Soc. Am. 116, 2365-2373 (2004); J. Speech Lang. Hear. Res. 55, 779-790 (2012)], respectively. Significant clear speech effects were observed for all acoustic metrics, although not all measures changed for all vowels and considerable talker variability was observed. Mixed-effects analyses revealed that the contribution of duration and steady-state formant information to vowel intelligibility differed for the two listener groups. This outcome is consistent with earlier research suggesting that hearing loss, and possibly aging, alters the way acoustic cues are used for identifying vowels.

  14. Professional Roles and Responsibilities in Meeting the Needs of Children with Speech, Language and Communication Needs: Joint Working between Educational Psychologists and Speech and Language Therapists

    ERIC Educational Resources Information Center

    McConnellogue, Sheila

    2011-01-01

    There is a large population of children with speech, language and communication needs who have additional special educational needs (SEN). Whilst professional collaboration between education and health professionals is recommended to ensure an integrated delivery of statutory services for this population of children, formal frameworks should be…

  15. Animal Models of Speech and Vocal Communication Deficits Associated With Psychiatric Disorders.

    PubMed

    Konopka, Genevieve; Roberts, Todd F

    2016-01-01

    Disruptions in speech, language, and vocal communication are hallmarks of several neuropsychiatric disorders, most notably autism spectrum disorders. Historically, the use of animal models to dissect molecular pathways and connect them to behavioral endophenotypes in cognitive disorders has proven to be an effective approach for developing and testing disease-relevant therapeutics. The unique aspects of human language compared with vocal behaviors in other animals make such an approach potentially more challenging. However, the study of vocal learning in species with analogous brain circuits to humans may provide entry points for understanding this human-specific phenotype and diseases. We review animal models of vocal learning and vocal communication and specifically link phenotypes of psychiatric disorders to relevant model systems. Evolutionary constraints in the organization of neural circuits and synaptic plasticity result in similarities in the brain mechanisms for vocal learning and vocal communication. Comparative approaches and careful consideration of the behavioral limitations among different animal models can provide critical avenues for dissecting the molecular pathways underlying cognitive disorders that disrupt speech, language, and vocal communication.

  16. Free Speech Yearbook: 1972.

    ERIC Educational Resources Information Center

    Tedford, Thomas L., Ed.

    This book is a collection of essays on free speech issues and attitudes, compiled by the Commission on Freedom of Speech of the Speech Communication Association. Four articles focus on freedom of speech in classroom situations as follows: a philosophic view of teaching free speech, effects of a course on free speech on student attitudes,…

  17. Acoustic Variations in Adductor Spasmodic Dysphonia as a Function of Speech Task.

    ERIC Educational Resources Information Center

    Sapienza, Christine M.; Walton, Suzanne; Murry, Thomas

    1999-01-01

    Acoustic phonatory events were identified in 14 women diagnosed with adductor spasmodic dysphonia (ADSD), a focal laryngeal dystonia that disturbs phonatory function, and compared with those of 14 age-matched women with no vocal dysfunction. Findings indicated ADSD subjects produced more aberrant acoustic events than controls during tasks of…

  18. Speech and Language Therapists' Approaches to Communication Intervention with Children and Adults with Profound and Multiple Learning Disability

    ERIC Educational Resources Information Center

    Goldbart, Juliet; Chadwick, Darren; Buell, Susan

    2014-01-01

    Background: People with profound intellectual and multiple disabilities (PMLD) have communication impairments as one defining characteristic. Aims: To explore speech and language therapists' (SLTs) decision making in communication interventions for people with PMLD, in terms of the intervention approaches used, the factors informing the…

  19. Coordination of Gaze and Speech in Communication between Children with Hearing Impairment and Normal-Hearing Peers

    ERIC Educational Resources Information Center

    Sandgren, Olof; Andersson, Richard; van de Weijer, Joost; Hansson, Kristina; Sahlén, Birgitta

    2014-01-01

    Purpose: To investigate gaze behavior during communication between children with hearing impairment (HI) and normal-hearing (NH) peers. Method: Ten HI-NH and 10 NH-NH dyads performed a referential communication task requiring description of faces. During task performance, eye movements and speech were tracked. Using verbal event (questions,…

  20. A Bioecological Framework to Evaluate Communicative Participation Outcomes for Preschoolers Receiving Speech-Language Therapy Interventions in Ontario, Canada

    ERIC Educational Resources Information Center

    Cunningham, Barbara J.; Rosenbaum, Peter L.

    2015-01-01

    Background: The Preschool Speech and Language Program (PSLP) in Ontario, Canada, is a publicly funded intervention service for children from birth to 5 years with communication disorders. It has begun a population-level programme evaluation of children's communicative participation outcomes following therapy. Data are currently being collected for…

  1. "The Communication Needs and Rights of Mankind", Group 1 Report of the Futuristic Priorities Division of the Speech Communication Association. "Future Communication Technologies; Hardware and Software"; Group 2 Report.

    ERIC Educational Resources Information Center

    Dance, Frank E. X.; And Others

    This paper reports on the Futuristic Priorities Division members' recommendations and priorities concerning the impact of the future on communication and on the speech communication discipline. The recommendations and priorities are listed for two subgroups: The Communication Needs and Rights of Mankind; and Future Communication Technologies:…

  2. Speech Communication and Communication Processes: Abstracts of Doctoral Dissertations Published in "Dissertation Abstracts International," April and May 1978 (Vol. 38 Nos. 10 and 11).

    ERIC Educational Resources Information Center

    ERIC Clearinghouse on Reading and Communication Skills, Urbana, IL.

    This collection of abstracts is part of a continuing series providing information on recent doctoral dissertations. The 25 titles deal with a variety of topics, including the following: the nature of creativity in advertising communication; speech communication difficulties of international professors; rhetorical arguments regarding the…

  3. The Body That Speaks: Recombining Bodies and Speech Sources in Unscripted Face-to-Face Communication

    PubMed Central

    Gillespie, Alex; Corti, Kevin

    2016-01-01

    This article examines advances in research methods that enable experimental substitution of the speaking body in unscripted face-to-face communication. A taxonomy of six hybrid social agents is presented by combining three types of bodies (mechanical, virtual, and human) with either an artificial or human speech source. Our contribution is to introduce and explore the significance of two particular hybrids: (1) the cyranoid method that enables humans to converse face-to-face through the medium of another person's body, and (2) the echoborg method that enables artificial intelligence to converse face-to-face through the medium of a human body. These two methods are distinct in being able to parse the unique influence of the human body when combined with various speech sources. We also introduce a new framework for conceptualizing the body's role in communication, distinguishing three levels: self's perspective on the body, other's perspective on the body, and self's perspective of other's perspective on the body. Within each level the cyranoid and echoborg methodologies make important research questions tractable. By conceptualizing and synthesizing these methods, we outline a novel paradigm of research on the role of the body in unscripted face-to-face communication. PMID:27660616

  4. The Body That Speaks: Recombining Bodies and Speech Sources in Unscripted Face-to-Face Communication.

    PubMed

    Gillespie, Alex; Corti, Kevin

    2016-01-01

    This article examines advances in research methods that enable experimental substitution of the speaking body in unscripted face-to-face communication. A taxonomy of six hybrid social agents is presented by combining three types of bodies (mechanical, virtual, and human) with either an artificial or human speech source. Our contribution is to introduce and explore the significance of two particular hybrids: (1) the cyranoid method that enables humans to converse face-to-face through the medium of another person's body, and (2) the echoborg method that enables artificial intelligence to converse face-to-face through the medium of a human body. These two methods are distinct in being able to parse the unique influence of the human body when combined with various speech sources. We also introduce a new framework for conceptualizing the body's role in communication, distinguishing three levels: self's perspective on the body, other's perspective on the body, and self's perspective of other's perspective on the body. Within each level the cyranoid and echoborg methodologies make important research questions tractable. By conceptualizing and synthesizing these methods, we outline a novel paradigm of research on the role of the body in unscripted face-to-face communication.

  5. The Body That Speaks: Recombining Bodies and Speech Sources in Unscripted Face-to-Face Communication.

    PubMed

    Gillespie, Alex; Corti, Kevin

    2016-01-01

    This article examines advances in research methods that enable experimental substitution of the speaking body in unscripted face-to-face communication. A taxonomy of six hybrid social agents is presented by combining three types of bodies (mechanical, virtual, and human) with either an artificial or human speech source. Our contribution is to introduce and explore the significance of two particular hybrids: (1) the cyranoid method that enables humans to converse face-to-face through the medium of another person's body, and (2) the echoborg method that enables artificial intelligence to converse face-to-face through the medium of a human body. These two methods are distinct in being able to parse the unique influence of the human body when combined with various speech sources. We also introduce a new framework for conceptualizing the body's role in communication, distinguishing three levels: self's perspective on the body, other's perspective on the body, and self's perspective of other's perspective on the body. Within each level the cyranoid and echoborg methodologies make important research questions tractable. By conceptualizing and synthesizing these methods, we outline a novel paradigm of research on the role of the body in unscripted face-to-face communication. PMID:27660616

  6. The Body That Speaks: Recombining Bodies and Speech Sources in Unscripted Face-to-Face Communication

    PubMed Central

    Gillespie, Alex; Corti, Kevin

    2016-01-01

    This article examines advances in research methods that enable experimental substitution of the speaking body in unscripted face-to-face communication. A taxonomy of six hybrid social agents is presented by combining three types of bodies (mechanical, virtual, and human) with either an artificial or human speech source. Our contribution is to introduce and explore the significance of two particular hybrids: (1) the cyranoid method that enables humans to converse face-to-face through the medium of another person's body, and (2) the echoborg method that enables artificial intelligence to converse face-to-face through the medium of a human body. These two methods are distinct in being able to parse the unique influence of the human body when combined with various speech sources. We also introduce a new framework for conceptualizing the body's role in communication, distinguishing three levels: self's perspective on the body, other's perspective on the body, and self's perspective of other's perspective on the body. Within each level the cyranoid and echoborg methodologies make important research questions tractable. By conceptualizing and synthesizing these methods, we outline a novel paradigm of research on the role of the body in unscripted face-to-face communication.

  7. Experimental studies of applications of time-reversal acoustics to noncoherent underwater communications

    NASA Astrophysics Data System (ADS)

    Heinemann, M.; Larraza, A.; Smith, K. B.

    2003-06-01

    The most difficult problem in shallow underwater acoustic communications is considered to be the time-varying multipath propagation because it impacts negatively on data rates. At high data rates the intersymbol interference requires adaptive algorithms on the receiver side that lead to computationally intensive and complex signal processing. A novel technique called time-reversal acoustics (TRA) can environmentally adapt the acoustic propagation effects of a complex medium in order to focus energy at a particular target range and depth. Using TRA, the multipath structure is reduced because all the propagation paths add coherently at the intended target location. This property of time-reversal acoustics suggests a potential application in the field of noncoherent acoustic communications. This work presents results of a tank scale experiment using an algorithm for rapid transmission of binary data in a complex underwater environment with the TRA approach. A simple 15-symbol code provides an example of the simplicity and feasibility of the approach. Covert coding due to the inherent scrambling induced by the environment at points other than the intended receiver is also investigated. The experiments described suggest a high potential in data rate for the time-reversal approach in underwater acoustic communications while keeping the computational complexity low.

  8. Acoustic Feature Optimization Based on F-Ratio for Robust Speech Recognition

    NASA Astrophysics Data System (ADS)

    Sun, Yanqing; Zhou, Yu; Zhao, Qingwei; Yan, Yonghong

    This paper focuses on the problem of performance degradation in mismatched speech recognition. The F-Ratio analysis method is utilized to analyze the significance of different frequency bands for speech unit classification, and we find that frequencies around 1kHz and 3kHz, which are the upper bounds of the first and the second formants for most of the vowels, should be emphasized in comparison to the Mel-frequency cepstral coefficients (MFCC). The analysis result is further observed to be stable in several typical mismatched situations. Similar to the Mel-Frequency scale, another frequency scale called the F-Ratio-scale is thus proposed to optimize the filter bank design for the MFCC features, and make each subband contains equal significance for speech unit classification. Under comparable conditions, with the modified features we get a relative 43.20% decrease compared with the MFCC in sentence error rate for the emotion affected speech recognition, 35.54%, 23.03% for the noisy speech recognition at 15dB and 0dB SNR (signal to noise ratio) respectively, and 64.50% for the three years' 863 test data. The application of the F-Ratio analysis on the clean training set of the Aurora2 database demonstrates its robustness over languages, texts and sampling rates.

  9. A high-frequency warm shallow water acoustic communications channel model and measurements.

    PubMed

    Chitre, Mandar

    2007-11-01

    Underwater acoustic communication is a core enabling technology with applications in ocean monitoring using remote sensors and autonomous underwater vehicles. One of the more challenging underwater acoustic communication channels is the medium-range very shallow warm-water channel, common in tropical coastal regions. This channel exhibits two key features-extensive time-varying multipath and high levels of non-Gaussian ambient noise due to snapping shrimp-both of which limit the performance of traditional communication techniques. A good understanding of the communications channel is key to the design of communication systems. It aids in the development of signal processing techniques as well as in the testing of the techniques via simulation. In this article, a physics-based channel model for the very shallow warm-water acoustic channel at high frequencies is developed, which are of interest to medium-range communication system developers. The model is based on ray acoustics and includes time-varying statistical effects as well as non-Gaussian ambient noise statistics observed during channel studies. The model is calibrated and its accuracy validated using measurements made at sea.

  10. Beyond ideal speech situations: adapting to communication asymmetries in health care.

    PubMed

    Gillespie, Alex; Reader, Tom; Cornish, Flora; Campbell, Catherine

    2014-01-01

    Inclusive, unconstrained and honest communication is widely advocated as beneficial and ethical. We critically explore this assumption by reflecting upon our research in acute care, informal care and public health. Using Habermas' ideals of dialogue to conceptualise ideal speech, we concur with observations that health care is often characterised by intractable exclusions and constraints. Rather than advocating implementing the ideals of dialogue, however, we examine how people adapt to these difficult and intransigent contexts. Non-ideal contexts, we find, sometimes call for non-ideal responses. Deception and furthering personal interests, and thus departing from the ideals of dialogue, can be adaptive responses.

  11. Measures to Evaluate the Effects of DBS on Speech Production

    PubMed Central

    Weismer, Gary; Yunusova, Yana; Bunton, Kate

    2011-01-01

    The purpose of this paper is to review and evaluate measures of speech production that could be used to document effects of Deep Brain Stimulation (DBS) on speech performance, especially in persons with Parkinson disease (PD). A small set of evaluative criteria for these measures is presented first, followed by consideration of several speech physiology and speech acoustic measures that have been studied frequently and reported on in the literature on normal speech production, and speech production affected by neuromotor disorders (dysarthria). Each measure is reviewed and evaluated against the evaluative criteria. Embedded within this review and evaluation is a presentation of new data relating speech motions to speech intelligibility measures in speakers with PD, amyotrophic lateral sclerosis (ALS), and control speakers (CS). These data are used to support the conclusion that at the present time the slope of second formant transitions (F2 slope), an acoustic measure, is well suited to make inferences to speech motion and to predict speech intelligibility. The use of other measures should not be ruled out, however, and we encourage further development of evaluative criteria for speech measures designed to probe the effects of DBS or any treatment with potential effects on speech production and communication skills. PMID:24932066

  12. Revisiting Neil Armstrongs Moon-Landing Quote: Implications for Speech Perception, Function Word Reduction, and Acoustic Ambiguity.

    PubMed

    Baese-Berk, Melissa M; Dilley, Laura C; Schmidt, Stephanie; Morrill, Tuuli H; Pitt, Mark A

    2016-01-01

    Neil Armstrong insisted that his quote upon landing on the moon was misheard, and that he had said one small step for a man, instead of one small step for man. What he said is unclear in part because function words like a can be reduced and spectrally indistinguishable from the preceding context. Therefore, their presence can be ambiguous, and they may disappear perceptually depending on the rate of surrounding speech. Two experiments are presented examining production and perception of reduced tokens of for and for a in spontaneous speech. Experiment 1 investigates the distributions of several acoustic features of for and for a. The results suggest that the distributions of for and for a overlap substantially, both in terms of temporal and spectral characteristics. Experiment 2 examines perception of these same tokens when the context speaking rate differs. The perceptibility of the function word a varies as a function of this context speaking rate. These results demonstrate that substantial ambiguity exists in the original quote from Armstrong, and that this ambiguity may be understood through context speaking rate. PMID:27603209

  13. Revisiting Neil Armstrongs Moon-Landing Quote: Implications for Speech Perception, Function Word Reduction, and Acoustic Ambiguity

    PubMed Central

    Baese-Berk, Melissa M.; Dilley, Laura C.; Schmidt, Stephanie; Morrill, Tuuli H.; Pitt, Mark A.

    2016-01-01

    Neil Armstrong insisted that his quote upon landing on the moon was misheard, and that he had said one small step for a man, instead of one small step for man. What he said is unclear in part because function words like a can be reduced and spectrally indistinguishable from the preceding context. Therefore, their presence can be ambiguous, and they may disappear perceptually depending on the rate of surrounding speech. Two experiments are presented examining production and perception of reduced tokens of for and for a in spontaneous speech. Experiment 1 investigates the distributions of several acoustic features of for and for a. The results suggest that the distributions of for and for a overlap substantially, both in terms of temporal and spectral characteristics. Experiment 2 examines perception of these same tokens when the context speaking rate differs. The perceptibility of the function word a varies as a function of this context speaking rate. These results demonstrate that substantial ambiguity exists in the original quote from Armstrong, and that this ambiguity may be understood through context speaking rate. PMID:27603209

  14. Revisiting Neil Armstrongs Moon-Landing Quote: Implications for Speech Perception, Function Word Reduction, and Acoustic Ambiguity.

    PubMed

    Baese-Berk, Melissa M; Dilley, Laura C; Schmidt, Stephanie; Morrill, Tuuli H; Pitt, Mark A

    2016-01-01

    Neil Armstrong insisted that his quote upon landing on the moon was misheard, and that he had said one small step for a man, instead of one small step for man. What he said is unclear in part because function words like a can be reduced and spectrally indistinguishable from the preceding context. Therefore, their presence can be ambiguous, and they may disappear perceptually depending on the rate of surrounding speech. Two experiments are presented examining production and perception of reduced tokens of for and for a in spontaneous speech. Experiment 1 investigates the distributions of several acoustic features of for and for a. The results suggest that the distributions of for and for a overlap substantially, both in terms of temporal and spectral characteristics. Experiment 2 examines perception of these same tokens when the context speaking rate differs. The perceptibility of the function word a varies as a function of this context speaking rate. These results demonstrate that substantial ambiguity exists in the original quote from Armstrong, and that this ambiguity may be understood through context speaking rate.

  15. Comparison of Acoustic and Kinematic Approaches to Measuring Utterance-Level Speech Variability

    ERIC Educational Resources Information Center

    Howell, Peter; Anderson, Andrew J.; Bartrip, Jon; Bailey, Eleanor

    2009-01-01

    Purpose: The spatiotemporal index (STI) is one measure of variability. As currently implemented, kinematic data are used, requiring equipment that cannot be used with some patient groups or in scanners. An experiment is reported that addressed whether STI can be extended to an audio measure of sound pressure of the speech envelope over time that…

  16. The Exploitation of Subphonemic Acoustic Detail in L2 Speech Segmentation

    ERIC Educational Resources Information Center

    Shoemaker, Ellenor

    2014-01-01

    The current study addresses an aspect of second language (L2) phonological acquisition that has received little attention to date--namely, the acquisition of allophonic variation as a word boundary cue. The role of subphonemic variation in the segmentation of speech by native speakers has been indisputably demonstrated; however, the acquisition of…

  17. Spectrally Efficient Underwater Acoustic Communications: Channel Characterization and Design Aspects for OFDM

    NASA Astrophysics Data System (ADS)

    Radosevic, Andreja

    In this dissertation, we consider design aspects of spectrally efficient underwater acoustic (UWA) communications. In particular, we first focus on statistical characterization and capacity evaluation of shallow water acoustic communications channels. Wideband single-carrier and multi-carrier probe signals are employed during the Kauai Acoustic Communications MURI 2008 (KAM08) and 2011 (KAM11) experiments, to measure the time-varying channel response, and to estimate its statistical properties and capacity that play an important role in the design of spectrally efficient communication systems. Besides the capacity analysis for unconstrained inputs, we determine new bounds on the achievable information rate for discrete-time Gaussian channels with inter-symbol interference and independent and uniformly distributed channel input symbols drawn from finite-order modulation alphabets. Specifically, we derived new bounds on the achievable rates for sparse channels with long memory. Furthermore, we explore design aspects of adaptive modulation based on orthogonal frequency division multiplexing (OFDM) for UWA communications, and study its performance using real-time at-sea experiments. Lastly, we investigate a channel estimation (CE) method for improving the spectral efficiency of UWA communications. Specifically, we determine the performance of a selective decision directed (DD) CE method for UWA OFDM-based communications.

  18. Electrophysiological and Kinematic Correlates of Communicative Intent in the Planning and Production of Pointing Gestures and Speech.

    PubMed

    Peeters, David; Chu, Mingyuan; Holler, Judith; Hagoort, Peter; Özyürek, Aslı

    2015-12-01

    In everyday human communication, we often express our communicative intentions by manually pointing out referents in the material world around us to an addressee, often in tight synchronization with referential speech. This study investigated whether and how the kinematic form of index finger pointing gestures is shaped by the gesturer's communicative intentions and how this is modulated by the presence of concurrently produced speech. Furthermore, we explored the neural mechanisms underpinning the planning of communicative pointing gestures and speech. Two experiments were carried out in which participants pointed at referents for an addressee while the informativeness of their gestures and speech was varied. Kinematic and electrophysiological data were recorded online. It was found that participants prolonged the duration of the stroke and poststroke hold phase of their gesture to be more communicative, in particular when the gesture was carrying the main informational burden in their multimodal utterance. Frontal and P300 effects in the ERPs suggested the importance of intentional and modality-independent attentional mechanisms during the planning phase of informative pointing gestures. These findings contribute to a better understanding of the complex interplay between action, attention, intention, and language in the production of pointing gestures, a communicative act core to human interaction.

  19. Performance analysis of a LDPC coded OFDM communication system in shallow water acoustic channels

    NASA Astrophysics Data System (ADS)

    Liu, Shengxing; Xu, Xiaomei

    2012-11-01

    Time-varying significant multipath interference is the major obstacle to reliable data communication in shallow water acoustic channels. In this paper, the performance of a low density parity check (LDPC) coded orthogonal frequency division multiplexing (OFDM) communication system is investigated for these channels. The initial message for LDPC, decoded by using the belief propagation (BP) algorithm, is deduced for OFDM underwater acoustic channels; based on this deduction, the noise thresholds of regular LDPC codes with different code rates are obtained by using the density evolution algorithm. Furthermore, a communication system model, developed with LDPC code, OFDM and channel interleaver for shallow water acoustic channels, is introduced. The effect of modulation and coding schemes on the LDPC codes performance is investigated by simulation. The results show that the system can achieve remarkable performance in shallow water acoustic channels, and the performance improves with increasing code length and decreasing code rate. The bit error rate (BER) of the system, under conditions with QPSK modulation, 1280-code length and 1/2-code rate, is less than 10-5 when the signal to noise ratio (SNR) is greater than 6.8dB. These values are obtained for a five-path shallow water acoustic channel of Xiamen harbor.

  20. Integrating Music Therapy Services and Speech-Language Therapy Services for Children with Severe Communication Impairments: A Co-Treatment Model

    ERIC Educational Resources Information Center

    Geist, Kamile; McCarthy, John; Rodgers-Smith, Amy; Porter, Jessica

    2008-01-01

    Documenting how music therapy can be integrated with speech-language therapy services for children with communication delay is not evident in the literature. In this article, a collaborative model with procedures, experiences, and communication outcomes of integrating music therapy with the existing speech-language services is given. Using…

  1. The Impact of Interrupted Use of a Speech Generating Device on the Communication Acts of a Child with Autism Spectrum Disorder: A Case Study

    ERIC Educational Resources Information Center

    Neeley, Richard A.; Pulliam, Mary Hannah; Catt, Merrill; McDaniel, D. Mike

    2015-01-01

    This case study examined the initial and renewed impact of speech generating devices on the expressive communication behaviors of a child with autism spectrum disorder. The study spanned six years of interrupted use of two speech generating devices. The child's communication behaviors were analyzed from video recordings and included communication…

  2. MOOD STATE PREDICTION FROM SPEECH OF VARYING ACOUSTIC QUALITY FOR INDIVIDUALS WITH BIPOLAR DISORDER

    PubMed Central

    Gideon, John; Provost, Emily Mower; McInnis, Melvin

    2016-01-01

    Speech contains patterns that can be altered by the mood of an individual. There is an increasing focus on automated and distributed methods to collect and monitor speech from large groups of patients suffering from mental health disorders. However, as the scope of these collections increases, the variability in the data also increases. This variability is due in part to the range in the quality of the devices, which in turn affects the quality of the recorded data, negatively impacting the accuracy of automatic assessment. It is necessary to mitigate variability effects in order to expand the impact of these technologies. This paper explores speech collected from phone recordings for analysis of mood in individuals with bipolar disorder. Two different phones with varying amounts of clipping, loudness, and noise are employed. We describe methodologies for use during preprocessing, feature extraction, and data modeling to correct these differences and make the devices more comparable. The results demonstrate that these pipeline modifications result in statistically significantly higher performance, which highlights the potential of distributed mental health systems. PMID:27570493

  3. Acoustic Analyses of Speech Sounds and Rhythms in Japanese- and English-Learning Infants

    PubMed Central

    Yamashita, Yuko; Nakajima, Yoshitaka; Ueda, Kazuo; Shimada, Yohko; Hirsh, David; Seno, Takeharu; Smith, Benjamin Alexander

    2013-01-01

    The purpose of this study was to explore developmental changes, in terms of spectral fluctuations and temporal periodicity with Japanese- and English-learning infants. Three age groups (15, 20, and 24 months) were selected, because infants diversify phonetic inventories with age. Natural speech of the infants was recorded. We utilized a critical-band-filter bank, which simulated the frequency resolution in adults’ auditory periphery. First, the correlations between the power fluctuations of the critical-band outputs represented by factor analysis were observed in order to see how the critical bands should be connected to each other, if a listener is to differentiate sounds in infants’ speech. In the following analysis, we analyzed the temporal fluctuations of factor scores by calculating autocorrelations. The present analysis identified three factors as had been observed in adult speech at 24 months of age in both linguistic environments. These three factors were shifted to a higher frequency range corresponding to the smaller vocal tract size of the infants. The results suggest that the vocal tract structures of the infants had developed to become adult-like configuration by 24 months of age in both language environments. The amount of utterances with periodic nature of shorter time increased with age in both environments. This trend was clearer in the Japanese environment. PMID:23450824

  4. Quantifying the Effect of Compression Hearing Aid Release Time on Speech Acoustics and Intelligibility

    ERIC Educational Resources Information Center

    Jenstad, Lorienne M.; Souza, Pamela E.

    2005-01-01

    Compression hearing aids have the inherent, and often adjustable, feature of release time from compression. Research to date does not provide a consensus on how to choose or set release time. The current study had 2 purposes: (a) a comprehensive evaluation of the acoustic effects of release time for a single-channel compression system in quiet and…

  5. Depression Diagnoses and Fundamental Frequency-Based Acoustic Cues in Maternal Infant-Directed Speech

    ERIC Educational Resources Information Center

    Porritt, Laura L.; Zinser, Michael C.; Bachorowski, Jo-Anne; Kaplan, Peter S.

    2014-01-01

    F[subscript 0]-based acoustic measures were extracted from a brief, sentence-final target word spoken during structured play interactions between mothers and their 3- to 14-month-old infants and were analyzed based on demographic variables and DSM-IV Axis-I clinical diagnoses and their common modifiers. F[subscript 0] range (?F[subscript 0]) was…

  6. Linguistic Emphasis in Maternal Speech to Preschool Language Learners with Language Impairments: An Acoustical Perspective.

    ERIC Educational Resources Information Center

    Scheffel, Debora L.; Ingrisano, Dennis R-S

    2000-01-01

    The nature of linguistic emphasis was studied from audio recordings of 29 mother-child dyads. Nineteen dyads involved mothers interacting with their 4-year-olds who evidenced language impairments. Approximately 84 percent of children evidencing language impairments could be so classified based on the acoustic variables associated with maternal use…

  7. Exploration of Acoustic Features for Automatic Vowel Discrimination in Spontaneous Speech

    ERIC Educational Resources Information Center

    Tyson, Na'im R.

    2012-01-01

    In an attempt to understand what acoustic/auditory feature sets motivated transcribers towards certain labeling decisions, I built machine learning models that were capable of discriminating between canonical and non-canonical vowels excised from the Buckeye Corpus. Specifically, I wanted to model when the dictionary form and the transcribed-form…

  8. Intelligibility of Telephone Speech for the Hearing Impaired When Various Microphones Are Used for Acoustic Coupling.

    ERIC Educational Resources Information Center

    Janota, Claus P.; Janota, Jeanette Olach

    1991-01-01

    Various candidate microphones were evaluated for acoustic coupling of hearing aids to a telephone receiver. Results from testing by 9 hearing-impaired adults found comparable listening performance with a pressure gradient microphone at a 10 decibel higher level of interfering noise than with a normal pressure-sensitive microphone. (Author/PB)

  9. Classifying acoustic signals into phoneme categories: average and dyslexic readers make use of complex dynamical patterns and multifractal scaling properties of the speech signal.

    PubMed

    Hasselman, Fred

    2015-01-01

    Several competing aetiologies of developmental dyslexia suggest that the problems with acquiring literacy skills are causally entailed by low-level auditory and/or speech perception processes. The purpose of this study is to evaluate the diverging claims about the specific deficient peceptual processes under conditions of strong inference. Theoretically relevant acoustic features were extracted from a set of artificial speech stimuli that lie on a /bAk/-/dAk/ continuum. The features were tested on their ability to enable a simple classifier (Quadratic Discriminant Analysis) to reproduce the observed classification performance of average and dyslexic readers in a speech perception experiment. The 'classical' features examined were based on component process accounts of developmental dyslexia such as the supposed deficit in Envelope Rise Time detection and the deficit in the detection of rapid changes in the distribution of energy in the frequency spectrum (formant transitions). Studies examining these temporal processing deficit hypotheses do not employ measures that quantify the temporal dynamics of stimuli. It is shown that measures based on quantification of the dynamics of complex, interaction-dominant systems (Recurrence Quantification Analysis and the multifractal spectrum) enable QDA to classify the stimuli almost identically as observed in dyslexic and average reading participants. It seems unlikely that participants used any of the features that are traditionally associated with accounts of (impaired) speech perception. The nature of the variables quantifying the temporal dynamics of the speech stimuli imply that the classification of speech stimuli cannot be regarded as a linear aggregate of component processes that each parse the acoustic signal independent of one another, as is assumed by the 'classical' aetiologies of developmental dyslexia. It is suggested that the results imply that the differences in speech perception performance between average and

  10. Classifying acoustic signals into phoneme categories: average and dyslexic readers make use of complex dynamical patterns and multifractal scaling properties of the speech signal.

    PubMed

    Hasselman, Fred

    2015-01-01

    Several competing aetiologies of developmental dyslexia suggest that the problems with acquiring literacy skills are causally entailed by low-level auditory and/or speech perception processes. The purpose of this study is to evaluate the diverging claims about the specific deficient peceptual processes under conditions of strong inference. Theoretically relevant acoustic features were extracted from a set of artificial speech stimuli that lie on a /bAk/-/dAk/ continuum. The features were tested on their ability to enable a simple classifier (Quadratic Discriminant Analysis) to reproduce the observed classification performance of average and dyslexic readers in a speech perception experiment. The 'classical' features examined were based on component process accounts of developmental dyslexia such as the supposed deficit in Envelope Rise Time detection and the deficit in the detection of rapid changes in the distribution of energy in the frequency spectrum (formant transitions). Studies examining these temporal processing deficit hypotheses do not employ measures that quantify the temporal dynamics of stimuli. It is shown that measures based on quantification of the dynamics of complex, interaction-dominant systems (Recurrence Quantification Analysis and the multifractal spectrum) enable QDA to classify the stimuli almost identically as observed in dyslexic and average reading participants. It seems unlikely that participants used any of the features that are traditionally associated with accounts of (impaired) speech perception. The nature of the variables quantifying the temporal dynamics of the speech stimuli imply that the classification of speech stimuli cannot be regarded as a linear aggregate of component processes that each parse the acoustic signal independent of one another, as is assumed by the 'classical' aetiologies of developmental dyslexia. It is suggested that the results imply that the differences in speech perception performance between average and

  11. Classifying acoustic signals into phoneme categories: average and dyslexic readers make use of complex dynamical patterns and multifractal scaling properties of the speech signal

    PubMed Central

    2015-01-01

    Several competing aetiologies of developmental dyslexia suggest that the problems with acquiring literacy skills are causally entailed by low-level auditory and/or speech perception processes. The purpose of this study is to evaluate the diverging claims about the specific deficient peceptual processes under conditions of strong inference. Theoretically relevant acoustic features were extracted from a set of artificial speech stimuli that lie on a /bAk/-/dAk/ continuum. The features were tested on their ability to enable a simple classifier (Quadratic Discriminant Analysis) to reproduce the observed classification performance of average and dyslexic readers in a speech perception experiment. The ‘classical’ features examined were based on component process accounts of developmental dyslexia such as the supposed deficit in Envelope Rise Time detection and the deficit in the detection of rapid changes in the distribution of energy in the frequency spectrum (formant transitions). Studies examining these temporal processing deficit hypotheses do not employ measures that quantify the temporal dynamics of stimuli. It is shown that measures based on quantification of the dynamics of complex, interaction-dominant systems (Recurrence Quantification Analysis and the multifractal spectrum) enable QDA to classify the stimuli almost identically as observed in dyslexic and average reading participants. It seems unlikely that participants used any of the features that are traditionally associated with accounts of (impaired) speech perception. The nature of the variables quantifying the temporal dynamics of the speech stimuli imply that the classification of speech stimuli cannot be regarded as a linear aggregate of component processes that each parse the acoustic signal independent of one another, as is assumed by the ‘classical’ aetiologies of developmental dyslexia. It is suggested that the results imply that the differences in speech perception performance between

  12. Speech, Speech!

    ERIC Educational Resources Information Center

    McComb, Gordon

    1982-01-01

    Discussion focuses on the nature of computer-generated speech and voice synthesis today. State-of-the-art devices for home computers are called text-to-speech (TTS) systems. Details about the operation and use of TTS synthesizers are provided, and the time saving in programing over previous methods is emphasized. (MP)

  13. Improving the speech intelligibility in classrooms

    NASA Astrophysics Data System (ADS)

    Lam, Choi Ling Coriolanus

    One of the major acoustical concerns in classrooms is the establishment of effective verbal communication between teachers and students. Non-optimal acoustical conditions, resulting in reduced verbal communication, can cause two main problems. First, they can lead to reduce learning efficiency. Second, they can also cause fatigue, stress, vocal strain and health problems, such as headaches and sore throats, among teachers who are forced to compensate for poor acoustical conditions by raising their voices. Besides, inadequate acoustical conditions can induce the usage of public address system. Improper usage of such amplifiers or loudspeakers can lead to impairment of students' hearing systems. The social costs of poor classroom acoustics will be large to impair the learning of children. This invisible problem has far reaching implications for learning, but is easily solved. Many researches have been carried out that they have accurately and concisely summarized the research findings on classrooms acoustics. Though, there is still a number of challenging questions remaining unanswered. Most objective indices for speech intelligibility are essentially based on studies of western languages. Even several studies of tonal languages as Mandarin have been conducted, there is much less on Cantonese. In this research, measurements have been done in unoccupied rooms to investigate the acoustical parameters and characteristics of the classrooms. The speech intelligibility tests, which based on English, Mandarin and Cantonese, and the survey were carried out on students aged from 5 years old to 22 years old. It aims to investigate the differences in intelligibility between English, Mandarin and Cantonese of the classrooms in Hong Kong. The significance on speech transmission index (STI) related to Phonetically Balanced (PB) word scores will further be developed. Together with developed empirical relationship between the speech intelligibility in classrooms with the variations

  14. Acoustic Communication in Fishes and Potential Effects of Noise.

    PubMed

    Mann, David A

    2016-01-01

    Many soniferous fishes such as cods and groupers are commercially important. Sounds are produced during courtship and spawning, and there is the potential for aquatic noise to interfere with critical behaviors and affect populations. There are few data on the response of wild populations of sound-producing fishes to acoustic noise. New motion and sound exposure fish tags could be used to assess the behavioral responses of large numbers of fish to noise exposure. Many factors, such as fishing mortality and environmental variability in prey supply, could also affect populations and potentially interact with the behavioral responses to noise.

  15. Acoustic Communication in Fishes and Potential Effects of Noise.

    PubMed

    Mann, David A

    2016-01-01

    Many soniferous fishes such as cods and groupers are commercially important. Sounds are produced during courtship and spawning, and there is the potential for aquatic noise to interfere with critical behaviors and affect populations. There are few data on the response of wild populations of sound-producing fishes to acoustic noise. New motion and sound exposure fish tags could be used to assess the behavioral responses of large numbers of fish to noise exposure. Many factors, such as fishing mortality and environmental variability in prey supply, could also affect populations and potentially interact with the behavioral responses to noise. PMID:26611018

  16. Using Others' Words: Conversational Use of Reported Speech by Individuals with Aphasia and Their Communication Partners.

    ERIC Educational Resources Information Center

    Hengst, Julie A.; Frame, Simone R.; Neuman-Stritzel, Tiffany; Gannaway, Rachel

    2005-01-01

    Reported speech, wherein one quotes or paraphrases the speech of another, has been studied extensively as a set of linguistic and discourse practices. Researchers agree that reported speech is pervasive, found across languages, and used in diverse contexts. However, to date, there have been no studies of the use of reported speech among…

  17. Combined spatial diversity and time equalization for broadband multiple channel underwater acoustic communications

    NASA Astrophysics Data System (ADS)

    Skoro Kaskarovska, Violeta

    High data rate acoustic communications become feasible with the use of communication systems that operate at high frequency. The high frequency acoustic transmission in shallow water endures severe distortion as a result of the extensive intersymbol interference and Doppler shift, caused by the time variable multipath nature of the channel. In this research a Single Input Multiple Output (SIMO) acoustic communication system is developed to improve the reliability of the high data rate communications at short range in the shallow water acoustic channel. The proposed SIMO communication system operates at very high frequency and combines spatial diversity and decision feedback equalizer in a multilevel adaptive configuration. The first configuration performs selective combining on the equalized signals from multiple receivers and generates quality feedback parameter for the next level of combining. The second configuration implements a form of turbo equalization to evaluate the individual receivers using the feedback parameters as decision symbols. The improved signals from individual receivers are used in the next iteration of selective combining. Multiple iterations are used to achieve optimal estimate of the received signal. The multilevel adaptive configuration is evaluated on experimental and simulated data using SIMO system with three, four and five receivers. The simulation channel model developed for this research is based on experimental channel and Rician fading channel model. The performance of the channel is evaluated in terms of Bit Error Rate (BER) and Signal-to-Noise-and-Interference Ratio (SNIR). Using experimental data with non-zero BER, multilevel adaptive spatial diversity can achieve BER of 0 % and SNIR gain of 3 dB. The simulation results show that the average BER and SNIR after multilevel combining improve dramatically compared to the single receiver, even in case of extremely high BER of individual received signals. The results demonstrate the

  18. Effects of ocean thermocline variability on noncoherent underwater acoustic communications.

    PubMed

    Siderius, Martin; Porter, Michael B; Hursky, Paul; McDonald, Vincent

    2007-04-01

    The performance of acoustic modems in the ocean is strongly affected by the ocean environment. A storm can drive up the ambient noise levels, eliminate a thermocline by wind mixing, and whip up violent waves and thereby break up the acoustic mirror formed by the ocean surface. The combined effects of these and other processes on modem performance are not well understood. The authors have been conducting experiments to study these environmental effects on various modulation schemes. Here the focus is on the role of the thermocline on a widely used modulation scheme (frequency-shift keying). Using data from a recent experiment conducted in 100-m-deep water off the coast of Kauai, HI, frequency-shift-key modulation performance is shown to be strongly affected by diurnal cycles in the thermocline. There is dramatic variation in performance (measured by bit error rates) between receivers in the surface duct and receivers in the thermocline. To interpret the performance variations in a quantitative way, a precise metric is introduced based on a signal-to-interference-noise ratio that encompasses both the ambient noise and intersymbol interference. Further, it will be shown that differences in the fading statistics for receivers in and out of the thermocline explain the differences in modem performance.

  19. Acoustic communication in Panthera tigris: A study of tiger vocalization and auditory receptivity

    NASA Astrophysics Data System (ADS)

    Walsh, Edward J.; Wang, Lily M.; Armstrong, Douglas L.; Curro, Thomas; Simmons, Lee G.; McGee, Joann

    2003-04-01

    Acoustic communication represents a primary mode of interaction within the sub-species of Panthera tigris and it is commonly known that their vocal repertoire consists of a relatively wide range of utterances that include roars, growls, grunts, hisses and chuffling, vocalizations that are in some cases produced with extraordinary power. P. tigris vocalizations are known to contain significant amounts of acoustic energy over a wide spectral range, with peak output occurring in a low frequency bandwidth in the case of roars. von Muggenthaler (2000) has also shown that roars and other vocal productions uttered by P. tigris contain energy in the infrasonic range. While it is reasonable to assume that low and infrasonic acoustic cues are used as communication signals among conspecifics in the wild, it is clearly necessary to demonstrate that members of the P. tigris sub-species are responsive to low and infrasonic acoustic signals. The auditory brainstem response has proven to be an effective tool in the characterization of auditory performance among tigers and the results of an ongoing study of both the acoustical properties of P. tigris vocalizations and their auditory receptivity support the supposition that tigers are not only responsive to low frequency stimulation, but exquisitely so.

  20. Examination of time-reversal acoustics in shallow water and applications to noncoherent underwater communications

    NASA Astrophysics Data System (ADS)

    Smith, Kevin B.; Abrantes, Antonio A. M.; Larraza, Andres

    2003-06-01

    The shallow water acoustic communication channel is characterized by strong signal degradation caused by multipath propagation and high spatial and temporal variability of the channel conditions. At the receiver, multipath propagation causes intersymbol interference and is considered the most important of the channel distortions. This paper examines the application of time-reversal acoustic (TRA) arrays, i.e., phase-conjugated arrays (PCAs), that generate a spatio-temporal focus of acoustic energy at the receiver location, eliminating distortions introduced by channel propagation. This technique is self-adaptive and automatically compensates for environmental effects and array imperfections without the need to explicitly characterize the environment. An attempt is made to characterize the influences of a PCA design on its focusing properties with particular attention given to applications in noncoherent underwater acoustic communication systems. Due to the PCA spatial diversity focusing properties, PC arrays may have an important role in an acoustic local area network. Each array is able to simultaneously transmit different messages that will focus only at the destination receiver node.

  1. Basic Research on Time-Reversal Waves in Deep Ocean for Long Acoustic Communication

    NASA Astrophysics Data System (ADS)

    Shimura, Takuya; Watanabe, Yoshitaka; Ochi, Hiroshi

    2005-06-01

    We have studied the focusing property of time-reversal waves and its application to acoustic communication in shallow water. In this study, this focusing property in the deep ocean and its application to long horizontal acoustic communication are discussed. The results are as follows. Even if a time-reversal array (TRA) does not expand from the sea surface to the seabed, time-reversal signals converge, and the focusing property is not significantly affected by the depths of the focus and TRA. Then, it is revealed that by using time reversal, it is possible to enssure communication channel over a long range, through the simulation. In phase modulation, time reversal can demodulate almost only by itself, while in amplitude and phase modulation, an adaptive filter to compensates further.

  2. Design of Low-power Wake-up Circuits in Underwater Acoustic Communication

    NASA Astrophysics Data System (ADS)

    Cuixia, Zhang; Jiaxin, Wu; Yuanxuan, Li

    In underwater acoustic communication, the power consumption of the underwater communication equipments at harsh conditions of marine environment is an important problem. Under that scenario, we propose a design of low-power wake-up circuits based on SCM C8051F020 system. Compare to traditional wake-up circuits which directly judge the energy of received signals, our approach can greatly reduce the misjudgment caused by the environmental disturbance, and the performance of energy conservation is effective. The low-power wake-up circuits possess a promising application prospect in the long-distance wireless underwater communication.

  3. Piezoelectric Vibrational and Acoustic Alert for a Personal Communication Device

    NASA Technical Reports Server (NTRS)

    Woodard, Stanley E. (Inventor); Hellbaum, Richard F. (Inventor); Daugherty, Robert H. (Inventor); Scholz, Raymond C. (Inventor); Little, Bruce D. (Inventor); Fox, Robert L. (Inventor); Denhardt, Gerald A. (Inventor); Jang, SeGon (Inventor); Balein, Rizza (Inventor)

    2001-01-01

    An alert apparatus for a personal communication device includes a mechanically prestressed piezoelectric wafer positioned within the personal communication device and an alternating voltage input line coupled at two points of the wafer where polarity is recognized. The alert apparatus also includes a variable frequency device coupled to the alternating voltage input line, operative to switch the alternating voltage on the alternating voltage input line at least between an alternating voltage having a first frequency and an alternating voltage having a second frequency. The first frequency is preferably sufficiently high so as to cause the wafer to vibrate at a resulting frequency that produces a sound perceptible by a human ear, and the second frequency is preferably sufficiently low so as to cause the wafer to vibrate at a resulting frequency that produces a vibration readily felt by a holder of the personal communication device.

  4. Using the picture exchange communication system (PECS) with children with autism: assessment of PECS acquisition, speech, social-communicative behavior, and problem behavior.

    PubMed Central

    Charlop-Christy, Marjorie H; Carpenter, Michael; Le, Loc; LeBlanc, Linda A; Kellet, Kristen

    2002-01-01

    The picture exchange communication system (PECS) is an augmentative communication system frequently used with children with autism (Bondy & Frost, 1994; Siegel, 2000; Yamall, 2000). Despite its common clinical use, no well-controlled empirical investigations have been conducted to test the effectiveness of PECS. Using a multiple baseline design, the present study examined the acquisition of PECS with 3 children with autism. In addition, the study examined the effects of PECS training on the emergence of speech in play and academic settings. Ancillary measures of social-communicative behaviors and problem behaviors were recorded. Results indicated that all 3 children met the learning criterion for PECS and showed concomitant increases in verbal speech. Ancillary gains were associated with increases in social-communicative behaviors and decreases in problem behaviors. The results are discussed in terms of the provision of empirical support for PECS as well as the concomitant positive side effects of its use. PMID:12365736

  5. An Investigation of the Relationships Between Personality and Types of Instructor Criticism in the Beginning Speech-Communication Course.

    ERIC Educational Resources Information Center

    Huntley, Jackson Richard

    The purpose of this study was to find a method for individualizing instruction in a beginning college-level speech communications course--a course which allows for student-instructor interaction. To determine the relationship between the personality typology of introversion-extraversion and types of oral instructor criticism, 76 students who had…

  6. What Should Be the Unique Application of Teaching in Departments of Speech Communication Located in Urban Environments?

    ERIC Educational Resources Information Center

    Galvin, Kathleen M.

    This paper focuses on certain approaches which an urban speech department can use as it contributes to the preparation of urban school teachers to communicate effectively with their students. The contents include: "Verbal and Nonverbal Codes," which discusses the teacher as an encoder of verbal messages and emphasizes that teachers must learn to…

  7. Behavioral and Physiological Responses to Child-Directed Speech as Predictors of Communication Outcomes in Children with Autism Spectrum Disorders

    ERIC Educational Resources Information Center

    Watson, Linda R.; Baranek, Grace T.; Roberts, Jane E.; David, Fabian J.; Perryman, Twyla Y.

    2010-01-01

    Purpose: To determine the extent to which behavioral and physiological responses during child-directed speech (CDS) correlate concurrently and predictively with communication skills in young children with autism spectrum disorders (ASD). Method: Twenty-two boys with ASD (initial mean age: 35 months) participated in a longitudinal study. At entry,…

  8. Teaching on the Foam of Discourse: Genealogical Perspectives on the Emergence of Speech and Communication as Academic Disciplines.

    ERIC Educational Resources Information Center

    Macke, Frank J.

    This essay examines and explicates the nature and pattern of epistemic transformations undergone by the United States academic discipline known as "speech communication." The system of examination and explication employed in the paper follows Michel Foucault's genealogical method of inquiry and, as such, offers a postmodern critical "history" of…

  9. Understanding Why Speech-Language Pathologists Rarely Pursue a PhD in Communication Sciences and Disorders

    ERIC Educational Resources Information Center

    Myotte, Theodore; Hutchins, Tiffany L.; Cannizzaro, Michael S.; Belin, Gayle

    2011-01-01

    Masters-level speech-language pathologists in communication sciences and disorders (n = 122) completed a survey soliciting their reasons for not pursuing doctoral study. Factor analysis revealed a four-factor solution including one reflecting a lack of interest in doctoral study (Factor 2) and one reflecting practical financial concerns (Factor…

  10. "Tramites" and "Transportes": The Acquisition of Second Language Communicative Competence for One Speech Event in Puno, Peru.

    ERIC Educational Resources Information Center

    Hornberger, Nancy H.

    1989-01-01

    Analyzes ethnographic data regarding one prolonged speech event, the negotiation of a driver's license at the Ministry of Transportation in Puno, Peru, from the perspective of Hymes' redefinition of linguistic competence. Implications for the acquisition of second language communicative competence are also discussed. (Author/CB)

  11. Increasing Early Childhood Educators' Use of Communication-Facilitating and Language-Modelling Strategies: Brief Speech and Language Therapy Training

    ERIC Educational Resources Information Center

    McDonald, David; Proctor, Penny; Gill, Wendy; Heaven, Sue; Marr, Jane; Young, Jane

    2015-01-01

    Intensive Speech and Language Therapy (SLT) training courses for Early Childhood Educators (ECEs) can have a positive effect on their use of interaction strategies that support children's communication skills. The impact of brief SLT training courses is not yet clearly understood. The aims of these two studies were to assess the impact of a brief…

  12. The Picture Exchange Communication System: Effects on Manding and Speech Development for School-Aged Children with Autism

    ERIC Educational Resources Information Center

    Tincani, Matt; Crozier, Shannon; Alazett, Shannon

    2006-01-01

    We examined the effects of the Picture Exchange Communication System (PECS; Frost & Bondy, 2002) on the manding (requesting) and speech development of school-aged children with autism. In study 1, two participants, Damian and Bob, were taught PECS within a delayed multiple baseline design. Both participants demonstrated increased levels of manding…

  13. Communication, Listening, Cognitive and Speech Perception Skills in Children with Auditory Processing Disorder (APD) or Specific Language Impairment (SLI)

    ERIC Educational Resources Information Center

    Ferguson, Melanie A.; Hall, Rebecca L.; Riley, Alison; Moore, David R.

    2011-01-01

    Purpose: Parental reports of communication, listening, and behavior in children receiving a clinical diagnosis of specific language impairment (SLI) or auditory processing disorder (APD) were compared with direct tests of intelligence, memory, language, phonology, literacy, and speech intelligibility. The primary aim was to identify whether there…

  14. Workplace Learning Curriculum Guides. Volume VIII: Enhanced Basic Skills--Listening Skills, Communications, Speech, Self-Esteem, Individual Workplace Skills.

    ERIC Educational Resources Information Center

    Colorado Community Coll. and Occupational Education System, Denver.

    This volume, the last of a series of eight curriculum guides compiled by the Colorado Workplace Learning Initiative: 1991-92, contains 11 workplace literacy courses on enhanced basic skills including listening skills, communications, speech, self-esteem, and individual workplace skills. Introductory materials include a table of contents and a list…

  15. Acoustic Communication in Butterflyfishes: Anatomical Novelties, Physiology, Evolution, and Behavioral Ecology.

    PubMed

    Tricas, Timothy C; Webb, Jacqueline F

    2016-01-01

    Coral reef fishes live in noisy environments that may challenge their capacity for acoustic communication. Butterflyfishes (Family Chaetodontidae) are prominent and ecologically diverse members of coral reef communities worldwide. The discovery of a novel association of anterior swim bladder horns with the lateral line canal system in the genus Chaetodon (the laterophysic connection) revealed a putative adaptation for enhancement of sound reception by the lateral line system and/or the ear. Behavioral studies show that acoustic communication is an important component of butterflyfish social behavior. All bannerfish (Forcipiger, Heniochus, and Hemitaurichthys) and Chaetodon species studied thus far produce several sound types at frequencies of <1 to >1000 Hz. Ancestral character state analyses predict the existence of both shared (head bob) and divergent (tail slap) acoustic behaviors in these two clades. Experimental auditory physiology shows that butterflyfishes are primarily sensitive to stimuli associated with hydrodynamic particle accelerations of ≤500 Hz. In addition, the gas-filled swim bladder horns in Chaetodon are stimulated by sound pressure, which enhances and extends their auditory sensitivity to 1700-2000 Hz. The broadband spectrum of ambient noise present on coral reefs overlaps with the frequency characteristics of their sounds, thus both the close social affiliations common among butterflyfishes and the evolution of the swim bladder horns in Chaetodon facilitate their short-range acoustic communication. Butterflyfishes provide a unique and unexpected opportunity to carry out studies of fish bioacoustics in the lab and the field that integrate the study of sensory anatomy, physiology, evolution, and behavioral ecology.

  16. Revisiting speech interference in classrooms.

    PubMed

    Picard, M; Bradley, J S

    2001-01-01

    A review of the effects of ambient noise and reverberation on speech intelligibility in classrooms has been completed because of the long-standing lack of agreement on preferred acoustical criteria for unconstrained speech accessibility and communication in educational facilities. An overwhelming body of evidence has been collected to suggest that noise levels in particular are usually far in excess of any reasonable prescription for optimal conditions for understanding speech in classrooms. Quite surprisingly, poor classroom acoustics seem to be the prevailing condition for both normally-hearing and hearing-impaired students with reported A-weighted ambient noise levels 4-37 dB above values currently agreed upon to provide optimal understanding. Revision of currently proposed room acoustic performance criteria to ensure speech accessibility for all students indicates the need for a guideline weighted for age and one for more vulnerable groups. For teens (12-year-olds and older) and young adults having normal speech processing in noise, ambient noise levels not exceeding 40 dBA are suggested as acceptable, and reverberation times of about 0.5 s are concluded to be optimum. Younger students, having normal speech processing in noise for their age, would require noise levels ranging from 39 dBA for 10-11-year-olds to only 28.5 dBA for 6-7-year-olds. By contrast, groups suspected of delayed speech processing in noise may require levels as low as only 21.5 dBA at age 6-7. As one would expect, these more vulnerable students would include the hearing-impaired in the course of language development and non-native listeners. PMID:11688542

  17. Development and evaluation of the environment and communication assessment toolkit with speech-language pathologists.

    PubMed

    Bruce, Carrie; Brush, Jennifer A; Sanford, Jon A; Calkins, Margaret P

    2013-04-01

    Communication dysfunction that results from dementia can be exacerbated by environmental barriers such as inadequate lighting, noisy conditions, poor or absent environmental cues, and visual clutter. Speech-language pathologists (SLPs) should address these environmental barriers as part of a comprehensive treatment plan for clients with dementia. The Environment and Communication Assessment Toolkit for Dementia Care (ECAT) was evaluated by SLPs to determine: (1) changes in awareness of environmental factors prior to and after training; (2) impact of the ECAT on practice as measured by changes in the number of environmental modifications recommended and made prior to and after training; (3) utility of the information as measured by the helpfulness, amount of new information, and usefulness of the ECAT; and (4) usability of the ECAT materials based on ease of use. The SLPs used the ECAT with clients with dementia who had functional limitations and required substantial assistance with daily activities. Results indicate that the ECAT is an effective tool for SLPs, providing information about the impact of the environment on communication and supplying sufficient resources to make recommendations and implement effective interventions. The ECAT successfully increased awareness of environmental modifications, influenced the practice of recommending environmental modifications, and had utility in diverse aspects of clinical practice.

  18. Predicting Speech Intelligibility with a Multiple Speech Subsystems Approach in Children with Cerebral Palsy

    ERIC Educational Resources Information Center

    Lee, Jimin; Hustad, Katherine C.; Weismer, Gary

    2014-01-01

    Purpose: Speech acoustic characteristics of children with cerebral palsy (CP) were examined with a multiple speech subsystems approach; speech intelligibility was evaluated using a prediction model in which acoustic measures were selected to represent three speech subsystems. Method: Nine acoustic variables reflecting different subsystems, and…

  19. Behaviours Associated with Acoustic Communication in Nile Tilapia (Oreochromis niloticus)

    PubMed Central

    Longrie, Nicolas; Poncin, Pascal; Denoël, Mathieu; Gennotte, Vincent; Delcourt, Johann; Parmentier, Eric

    2013-01-01

    Background Sound production is widespread among fishes and accompanies many social interactions. The literature reports twenty-nine cichlid species known to produce sounds during aggressive and courtship displays, but the precise range in behavioural contexts is unclear. This study aims to describe the various Oreochromis niloticus behaviours that are associated with sound production in order to delimit the role of sound during different activities, including agonistic behaviours, pit activities, and reproduction and parental care by males and females of the species. Methodology/Principal Findings Sounds mostly occur during the day. The sounds recorded during this study accompany previously known behaviours, and no particular behaviour is systematically associated with sound production. Males and females make sounds during territorial defence but not during courtship and mating. Sounds support visual behaviours but are not used alone. During agonistic interactions, a calling Oreochromis niloticus does not bite after producing sounds, and more sounds are produced in defence of territory than for dominating individuals. Females produce sounds to defend eggs but not larvae. Conclusion/Significance Sounds are produced to reinforce visual behaviours. Moreover, comparisons with O. mossambicus indicate two sister species can differ in their use of sound, their acoustic characteristics, and the function of sound production. These findings support the role of sounds in differentiating species and promoting speciation. They also make clear that the association of sounds with specific life-cycle roles cannot be generalized to the entire taxa. PMID:23620756

  20. Channel Equalization for Single Carrier MIMO Underwater Acoustic Communications

    NASA Astrophysics Data System (ADS)

    Tao, Jun; Zheng, Yahong Rosa; Xiao, Chengshan; Yang, T. C.; Yang, Wen-Bin

    2010-12-01

    Multiple-input multiple-output (MIMO) underwater acoustic (UWA) channels introduce both space-time interference (STI) and time-varying phase distortion for transmitted signals. In such cases, the equalized symbols produced by conventional equalizer aiming for STI cancelation suffer phase rotation and thus cannot be reliably detected. In this paper, we propose a new equalization scheme for high data rate single carrier MIMO UWA channels. Different from existing methods employing joint equalization and symbolwise phase tracking technology, the proposed scheme decouples the interference cancelation (IC) operation and the phase compensation operation, leading to a generalized equalizer structure combining an IC equalizer with a phase compensator. The decoupling of the two functionalities leads to robust signal detection, which is most desirable in practical UWA applications. MIMO linear equalizer (LE) is adopted to remove space-time interference, and a groupwise phase estimation and correction method is used to compensate the phase rotation. In addition, the layered space-time processing technology is adopted to enhance the equalization performance. The proposed equalization scheme is tested to be very robust with extensive experimental data collected at Kauai, Hawaii, in September 2005, and Saint Margaret's Bay, Nova Scotia, Canada, in May 2006.

  1. [Quantification and improvement of speech transmission performance using headphones in acoustic stimulated functional magnetic resonance imaging].

    PubMed

    Yamamura, Ken ichiro; Takatsu, Yasuo; Miyati, Tosiaki; Kimura, Tetsuya

    2014-10-01

    Functional magnetic resonance imaging (fMRI) has made a major contribution to the understanding of higher brain function, but fMRI with auditory stimulation, used in the planning of brain tumor surgery, is often inaccurate because there is a risk that the sounds used in the trial may not be correctly transmitted to the subjects due to acoustic noise. This prompted us to devise a method of digitizing sound transmission ability from the accuracy rate of 67 syllables, classified into three types. We evaluated this with and without acoustic noise during imaging. We also improved the structure of the headphones and compared their sound transmission ability with that of conventional headphones attached to an MRI device (a GE Signa HDxt 3.0 T). We calculated and compared the sound transmission ability of the conventional headphones with that of the improved model. The 95 percent upper confidence limit (UCL) was used as the threshold for accuracy rate of hearing for both headphone models. There was a statistically significant difference between the conventional model and the improved model during imaging (p < 0.01). The rate of accuracy of the improved model was 16 percent higher. 29 and 22 syllables were accurate at a 95% UCL in the improved model and the conventional model, respectively. This study revealed the evaluation system used in this study to be useful for correctly identifying syllables during fMRI.

  2. Effects of speech style, room acoustics, and vocal fatigue on vocal effort.

    PubMed

    Bottalico, Pasquale; Graetzer, Simone; Hunter, Eric J

    2016-05-01

    Vocal effort is a physiological measure that accounts for changes in voice production as vocal loading increases. It has been quantified in terms of sound pressure level (SPL). This study investigates how vocal effort is affected by speaking style, room acoustics, and short-term vocal fatigue. Twenty subjects were recorded while reading a text at normal and loud volumes in anechoic, semi-reverberant, and reverberant rooms in the presence of classroom babble noise. The acoustics in each environment were modified by creating a strong first reflection in the talker position. After each task, the subjects answered questions addressing their perception of the vocal effort, comfort, control, and clarity of their own voice. Variation in SPL for each subject was measured per task. It was found that SPL and self-reported effort increased in the loud style and decreased when the reflective panels were present and when reverberation time increased. Self-reported comfort and control decreased in the loud style, while self-reported clarity increased when panels were present. The lowest magnitude of vocal fatigue was experienced in the semi-reverberant room. The results indicate that early reflections may be used to reduce vocal effort without modifying reverberation time.

  3. Acoustic evidence for the development of gestural coordination in the speech of 2-year-olds: a longitudinal study.

    PubMed

    Goodell, E W; Studdert-Kennedy, M

    1993-08-01

    Studies of child phonology have often assumed that young children first master a repertoire of phonemes and then build their lexicon by forming combinations of these abstract, contrastive units. However, evidence from children's systematic errors suggests that children first build a repertoire of words as integral sequences of gestures and then gradually differentiate these sequences into their gestural and segmental components. Recently, experimental support for this position has been found in the acoustic records of the speech of 3-, 5-, and 7-year-old children, suggesting that even in older children some phonemes have not yet fully segregated as units of gestural organization and control. The present longitudinal study extends this work to younger children (22- and 32-month-olds). Results demonstrate clear differences in the duration and coordination of gestures between children and adults, and a clear shift toward the patterns of adult speakers during roughly the third year of life. Details of the child-adult differences and developmental changes vary from one aspect of an utterance to another. PMID:8377484

  4. Single- and multi-channel underwater acoustic communication channel capacity: a computational study.

    PubMed

    Hayward, Thomas J; Yang, T C

    2007-09-01

    Acoustic communication channel capacity determines the maximum data rate that can be supported by an acoustic channel for a given source power and source/receiver configuration. In this paper, broadband acoustic propagation modeling is applied to estimate the channel capacity for a time-invariant shallow-water waveguide for a single source-receiver pair and for vertical source and receiver arrays. Without bandwidth constraints, estimated single-input, single-output (SISO) capacities approach 10 megabitss at 1 km range, but beyond 2 km range they decay at a rate consistent with previous estimates by Peloquin and Leinhos (unpublished, 1997), which were based on a sonar equation calculation. Channel capacities subject to source bandwidth constraints are approximately 30-90% lower than for the unconstrained case, and exhibit a significant wind speed dependence. Channel capacity is investigated for single-input, multi-output (SIMO) and multi-input, multi-output (MIMO) systems, both for finite arrays and in the limit of a dense array spanning the entire water column. The limiting values of the SIMO and MIMO channel capacities for the modeled environment are found to be about four times higher and up to 200-400 times higher, respectively, than for the SISO case. Implications for underwater acoustic communication systems are discussed.

  5. Subjective evaluation of speech and noise in learning environments in the realm of classroom acoustics: Results from laboratory and field experiments

    NASA Astrophysics Data System (ADS)

    Meis, Markus; Nocke, Christian; Hofmann, Simone; Becker, Bernhard

    2005-04-01

    The impact of different acoustical conditions in learning environments on noise annoyance and the evaluation of speech quality were tested in a series of three experiments. In Experiment 1 (n=79) the auralization of seven classrooms with reverberation times from 0.55 to 3.21 s [average between 250 Hz to 2 kHz] served to develop a Semantic Differential, evaluating a simulated teacher's voice. Four factors were found: acoustical comfort, roughness, sharpness, and loudness. In Experiment 2, the effects of two classroom renovations were examined from a holistic perspective. The rooms were treated acoustically with acoustic ceilings (RT=0.5 s [250 Hz-2 kHz]) and muffling floor materials as well as non-acoustically with a new lighting system and color design. The results indicate that pupils (n=61) in renovated classrooms judged the simulated voice more positively, were less annoyed from the noise in classrooms, and were more motivated to participate in the lessons. In Experiment 3 the sound environments from six different lecture rooms (RT=0.8 to 1.39 s [250 Hz-2 kHz]) in two Universities of Oldenburg were evaluated by 321 students during the lectures. Evidence found supports the assumption that acoustical comfort in rooms is dependent on frequency for rooms with higher reverberation times.

  6. Intonation and Communicative Intent in Mothers' Speech to Infants: Is the Melody the Message?

    ERIC Educational Resources Information Center

    Fernald, Anne

    1989-01-01

    Explored the power of intonation of speech addressed to adults and preverbal infants to convey meaningful information to 80 adult listeners. Listeners used intonation to identify speaker's intent with greater accuracy in infant-directed speech than adult-directed speech. (RJC)

  7. Diversity-based acoustic communication with a glider in deep water.

    PubMed

    Song, H C; Howe, Bruce M; Brown, Michael G; Andrew, Rex K

    2014-03-01

    The primary use of underwater gliders is to collect oceanographic data within the water column and periodically relay the data at the surface via a satellite connection. In summer 2006, a Seaglider equipped with an acoustic recording system received transmissions from a broadband acoustic source centered at 75 Hz deployed on the bottom off Kauai, Hawaii, while moving away from the source at ranges up to ∼200 km in deep water and diving up to 1000-m depth. The transmitted signal was an m-sequence that can be treated as a binary-phase shift-keying communication signal. In this letter multiple receptions are exploited (i.e., diversity combining) to demonstrate the feasibility of using the glider as a mobile communication gateway.

  8. Experimental study of an acoustic communication system in shallow-water conditions

    NASA Astrophysics Data System (ADS)

    Bobrovskii, I. V.; Yagotinets, V. P.

    2013-11-01

    The paper presents the results of experimental studies of an acoustic communication system containing parallel transmission digital information channels, which came about from applying the principles of orthogonal frequency division of individual information channels, resulting in an increased information rate. Received signals were processed by two separate receivers differing in the method by which the pulse response parameters of the medium were estimated during signal propagation. It is shown that to increase the reliability of received digital information for coherent separated reception of a multipath signal, it is necessary to estimate the parameters of this signal's components in frequency bands corresponding to individual information channels. When such estimates were taken into account, the acoustic communication system provided error-less reception of digital information in all experiments conducted under conditions of the shallow Volgograd Reservoir and Lake Ladoga.

  9. Symbolic Speech

    ERIC Educational Resources Information Center

    Podgor, Ellen S.

    1976-01-01

    The concept of symbolic speech emanates from the 1967 case of United States v. O'Brien. These discussions of flag desecration, grooming and dress codes, nude entertainment, buttons and badges, and musical expression show that the courts place symbolic speech in different strata from verbal communication. (LBH)

  10. Communication adjustment of patients with a laryngectomy in Turkey: analysis by type of surgery and mode of speech.

    PubMed

    Evitts, Paul M; Kasapoglu, Fikret; Demirci, Ugur; Miller, Julia Sterne

    2011-12-01

    Previous studies have routinely shown that individuals with a laryngectomy (IWL) have a decreased quality of life (QoL). However, the questionnaires frequently used are not specifically designed for persons with laryngeal cancer and may not reflect the issues of most concern to IWL. The purpose of this study was to investigate (1) the overall communication adjustment of IWL in Turkey, (2) the differences in communication adjustment among types of surgery for the treatment of laryngeal cancer, and (3) the differences in communication adjustment among types of speech following treatment for laryngeal cancer. Fifty-two patients (50 males and two females) who underwent surgical treatment for laryngeal cancer (i.e. total, supraglottic, frontolateral, vertical, supracricoid) and were being seen for follow-up care were recruited from an ENT clinic in Bursa, Turkey. Patients were provided with a written translation of the Self-Evaluation of Communication Experiences after Laryngectomy (SECEL) which yielded information on three subscales (General, Environment, and Attitude) and a Total score. Kruskall-Wallis and Wilcoxon signed ranks nonparametric tests of significance showed that overall adjustment was within the well-adjusted range for all types of surgery and modes of speech with two exceptions: Mean scores for supracricoid laryngectomy with cricohyoidopexy (28.5) and esophageal speech (29.3) were both within the poorly adjusted range on the Environment subscale [lowest mean score = 8.7 (supraglottic)]. Total mean scores ranged from 19.8 (supraglottic) to 49.9 (esophageal speech). Mean scores of the General subscale ranged from 4.0 (electrolaryngeal speech) to 7.7 (tracheoesophageal speech). Mean scores of the Attitude subscale ranged from 6.2 (supraglottic) to 19 (electrolaryngeal). Results of the Kruskall-Wallis test also showed a significant difference between type of surgery for the Environment subscale (p = 0.003), the Attitude subscale (p = 0.039), and the

  11. Underwater acoustic communication using orthogonal signal division multiplexing scheme with time diversity

    NASA Astrophysics Data System (ADS)

    Ebihara, Tadashi; Ogasawara, Hanako; Mizutani, Koichi

    2016-03-01

    In this paper, an underwater acoustic (UWA) communication scheme for mobile platforms is proposed. The proposed scheme is based on the orthogonal signal division multiplexing (OSDM) scheme, which offers highly reliable UWA communication. However, OSDM is not suitable for mobile platforms as it is — it requires a receiver array and a large calculation cost for equalization. To establish a reliable link with small communication platforms, we design OSDM that can perform reliable communication without the need for an array and can reduce receiver complexity using the time-diversity technique (TD), and evaluate its performance in experiments. The experimental results suggest that OSDM-TD can simultaneously achieve power-efficient communications and receiver complexity reduction, and can realize small-scale communication platforms. In detail, OSDM-TD achieved almost the same communication quality as conventional OSDM, in exchange for an effective data rate. Moreover, the power efficiency of OSDM-TD was almost the same as that of conventional OSDM with two receiver array elements, although the calculation cost of OSDM-TD was far below that of conventional OSDM. As a result, it was found that OSDM-TD is suitable for UWA communication for mobile nodes whose capacity and computational resources are severely limited.

  12. Multipath Effects on High-Frequency Coherent Acoustic Communications in Shallow Water

    NASA Astrophysics Data System (ADS)

    Son, Su-Uk; Kim, Hyeonsu; Joo, Jongmin; Choi, Jee Woong

    2013-07-01

    Shallow-water acoustic communication channel, referred to as a multipath-limited channel, produces inter-symbol interference that poses a significant obstacle to reliable communication. Accordingly, signal-to-multipath ratio (SMR), rather than signal-to-noise ratio (SNR), becomes an important factor affecting communication performance. However, it is difficult to estimate SMR from measured communication data, especially at higher frequency (>10 kHz) because many arrivals scattered from rough ocean boundaries produce a significant intrapath time spreading, which acts as random noise in communication. In this paper, the energy fraction of the channel impulse response existing in one symbol duration is proposed as a parameter for estimating the quality of shallow-water communication channels. This parameter is compared with the bit-error-rate performance for data acquired in shallow water off the south coast of Korea, where the water depth is 45 m and the bottom consists of sandy clay sediment. The results imply that the energy fraction in one symbol duration may be used as a parameter for describing shallow-water communication channels and applied to the quick decision of a symbol or bit rate in a shallow-water field for reliable underwater communication.

  13. Proximate and ultimate aspects of acoustic and multimodal communication in butterflyfishes

    NASA Astrophysics Data System (ADS)

    Boyle, Kelly S.

    Communication in social animals is shaped by natural selection on both sender and receiver. Diurnal butterflyfishes use a combination of visual cues like bright color patterns and motor pattern driven displays, acoustic communication, and olfactory cues that may advertise territorial behavior, facilitate recognition of individuals, and provide cues for courtship. This dissertation examines proximate and multimodal communication in several butterflyfishes, with an emphasis on acoustic communication which has recently garnered attention within the Chaetodontidae. Sound production in the genus Forcipiger involves a novel mechanism with synchronous contractions of opposing head muscles at the onset of sound emission and rapid cranial rotation that lags behind sound emission. Acoustic signals in F. flavissimus provide an accurate indicator of body size, and to a lesser extent cranial rotation velocity and acceleration. The closely related Hemitaurichthys polylepis produces rapid pulse trains of similar duration and spectral content to F. flavissimus, but with a dramatically different mechanism which involves contractions of hypaxial musculature at the anterior end of the swim bladder that occur with synchronous muscle action potentials. Both H. polylepis sonic and hypaxial trunk muscle fibers have triads at the z-line, but sonic fibers have smaller cross-sectional areas, more developed sarcoplasmic reticula, longer sarcomere lengths, and wider t-tubules. Sonic motor neurons are located along a long motor column entirely within the spinal cord and are composed of large and small types. Forcipiger flavissimus and F. longirostris are site attached and territorial, with F. flavissimus engaged in harem polygyny and F. longirostris in social monogamy. Both produce similar pulse sounds to conspecifics during territoriality that vary little with respect to communicative context. Chaetodon multicinctus can discriminate between mates and non-mate intruders, but require combined

  14. Wideband Multichannel Time-Reversal Processing for Acoustic Communications in a Tunnel-like Structure

    SciTech Connect

    Candy, J V; Chambers, D H; Robbins, C L; Guidry, B L; Poggio, A J; Dowla, F; Hertzog, C A

    2006-01-12

    The development of multichannel time-reversal (T/R) processing techniques continues to progress rapidly especially when the need to communicate in a highly reverberative environment becomes critical. The underlying T/R concept is based on time-reversing the Green's function characterizing the uncertain communications channel investigating the deleterious dispersion and multipath effects. In this paper, attention is focused on two major objectives: (1) wideband communications leading to a time reference modulation technique; and (2) multichannel acoustic communications in a tunnel (or cave or pipe) with many obstructions, multipath returns, severe background noise, disturbances, long propagation paths ({approx}180) with disruptions (bends). For this extremely hostile environment, it is shown that multichannel T/R receivers can easily be extended to the wideband designs while demonstrating their performance in both the ''canonical'' stairwell of our previous work as well as a tunnel-like structure. Acoustic information signals are transmitted with an 8-element host or base station array to two client receivers with a significant loss in signal levels due to the propagation environment. In this paper, the results of the new wideband T/R processor and modulation scheme are discussed to demonstrate the overall performance for both high (24-bit) and low (1-bit) bit level analog-to-digital (A/D) converter designs. These results are validated by performing proof-of-principle acoustic communications experiments in air. It is shown that the resulting T/R receivers are capable of extracting the transmitted coded sequence from noisy microphone array measurements with zero-bit error.

  15. Acoustic and aerodynamic characteristics of Country-Western, Operatic and Broadway singing styles compared to speech

    NASA Astrophysics Data System (ADS)

    Stone, Robert E.; Cleveland, Thomas F.; Sundberg, P. Johan

    2003-04-01

    Acoustic and aerodynamic measures were used to objectively describe characteristics of Country-Western (C-W) singing in a group of six premier performers in a series of studies and of operatic and Broadway singing in a female subject with professional experience in both styles of singing. For comparison purposes the same measures also were applied to individuals while speaking the same material as sung. Changes in pitch and vocal loudness were investigated for various dependent variables, including subglottal pressure, closed quotient, glottal leakage, H1-H2 difference [the level difference between the two lowest partials of the source spectrum and glottal compliance (the ratio between the air volume displaced in a glottal pulse and the subglottal pressure)], formant frequencies, long-term-average spectrum and vibrato characteristics (in operatic versus Broadway singing). Data from C-W singers suggest they use higher sub-glottal pressures in singing than in speaking. Changes in vocal intensity for doubling sub-glottal pressure is less than reported for classical singers. Several measures were similar for both speaking and C-W singing. Whereas results provide objective specification of differences between operatic and Broadway styles of singing, the latter seems similar to features of conversational speaking style.

  16. Time-Frequency Feature Representation Using Multi-Resolution Texture Analysis and Acoustic Activity Detector for Real-Life Speech Emotion Recognition

    PubMed Central

    Wang, Kun-Ching

    2015-01-01

    The classification of emotional speech is mostly considered in speech-related research on human-computer interaction (HCI). In this paper, the purpose is to present a novel feature extraction based on multi-resolutions texture image information (MRTII). The MRTII feature set is derived from multi-resolution texture analysis for characterization and classification of different emotions in a speech signal. The motivation is that we have to consider emotions have different intensity values in different frequency bands. In terms of human visual perceptual, the texture property on multi-resolution of emotional speech spectrogram should be a good feature set for emotion classification in speech. Furthermore, the multi-resolution analysis on texture can give a clearer discrimination between each emotion than uniform-resolution analysis on texture. In order to provide high accuracy of emotional discrimination especially in real-life, an acoustic activity detection (AAD) algorithm must be applied into the MRTII-based feature extraction. Considering the presence of many blended emotions in real life, in this paper make use of two corpora of naturally-occurring dialogs recorded in real-life call centers. Compared with the traditional Mel-scale Frequency Cepstral Coefficients (MFCC) and the state-of-the-art features, the MRTII features also can improve the correct classification rates of proposed systems among different language databases. Experimental results show that the proposed MRTII-based feature information inspired by human visual perception of the spectrogram image can provide significant classification for real-life emotional recognition in speech. PMID:25594590

  17. Feeling addressed! The role of body orientation and co-speech gesture in social communication.

    PubMed

    Nagels, Arne; Kircher, Tilo; Steines, Miriam; Straube, Benjamin

    2015-05-01

    During face-to-face communication, body orientation and coverbal gestures influence how information is conveyed. The neural pathways underpinning the comprehension of such nonverbal social cues in everyday interaction are to some part still unknown. During fMRI data acquisition, 37 participants were presented with video clips showing an actor speaking short sentences. The actor produced speech-associated iconic gestures (IC) or no gestures (NG) while he was visible either from an egocentric (ego) or from an allocentric (allo) position. Participants were asked to indicate via button press whether they felt addressed or not. We found a significant interaction of body orientation and gesture in addressment evaluations, indicating that participants evaluated IC-ego conditions as most addressing. The anterior cingulate cortex (ACC) and left fusiform gyrus were stronger activated for egocentric versus allocentric actor position in gesture context. Activation increase in the ACC for IC-ego>IC-allo further correlated positively with increased addressment ratings in the egocentric gesture condition. Gesture-related activation increase in the supplementary motor area, left inferior frontal gyrus and right insula correlated positively with gesture-related increase of addressment evaluations in the egocentric context. Results indicate that gesture use and body-orientation contribute to the feeling of being addressed and together influence neural processing in brain regions involved in motor simulation, empathy and mentalizing. PMID:25640962

  18. Feeling addressed! The role of body orientation and co-speech gesture in social communication.

    PubMed

    Nagels, Arne; Kircher, Tilo; Steines, Miriam; Straube, Benjamin

    2015-05-01

    During face-to-face communication, body orientation and coverbal gestures influence how information is conveyed. The neural pathways underpinning the comprehension of such nonverbal social cues in everyday interaction are to some part still unknown. During fMRI data acquisition, 37 participants were presented with video clips showing an actor speaking short sentences. The actor produced speech-associated iconic gestures (IC) or no gestures (NG) while he was visible either from an egocentric (ego) or from an allocentric (allo) position. Participants were asked to indicate via button press whether they felt addressed or not. We found a significant interaction of body orientation and gesture in addressment evaluations, indicating that participants evaluated IC-ego conditions as most addressing. The anterior cingulate cortex (ACC) and left fusiform gyrus were stronger activated for egocentric versus allocentric actor position in gesture context. Activation increase in the ACC for IC-ego>IC-allo further correlated positively with increased addressment ratings in the egocentric gesture condition. Gesture-related activation increase in the supplementary motor area, left inferior frontal gyrus and right insula correlated positively with gesture-related increase of addressment evaluations in the egocentric context. Results indicate that gesture use and body-orientation contribute to the feeling of being addressed and together influence neural processing in brain regions involved in motor simulation, empathy and mentalizing.

  19. Acoustics and its relation to language: The influence of Dick Bolt

    NASA Astrophysics Data System (ADS)

    Stevens, Kenneth N.

    2003-04-01

    Under the mentoring of Dick Bolt, and the stimulation he provided in the Acoustics Laboratory at MIT, many students were exposed to a range of topics in acoustics, including mechanisms of sound generation, radiation, and propagation, noise control, acoustics of resonators and rooms, human responses to sound, and speech perception under various adverse conditions. As someone who became interested in speech communication, I have recognized that this kind of quantitative background in acoustics is an important requirement for developing models of how humans produce speech, how they perceive and understand speech, and how children acquire these skills. Speech production involves sound sources produced by a nonlinear mechanical system and by noise arising from turbulent airflow. Sound is propagated in a vocal tract with yielding walls, and acoustic coupling is introduced by lossy resonators attached to the vocal tract, including the trachea and the nasal cavity. These acoustic principles of sound generation create an inventory of sound types that give rise to distinctive responses in the ears and brains of listeners. The solid grounding in acoustics provided by Dick Bolt and his leadership have helped in the formation of this linkage between acoustics, speech physiology, linguistics, and human perception.

  20. Opportunities in Speech Pathology.

    ERIC Educational Resources Information Center

    Newman, Parley W.

    The importance of speech is discussed and speech pathology is described. Types of communication disorders considered are articulation disorders, aphasia, facial deformity, hearing loss, stuttering, delayed speech, voice disorders, and cerebral palsy; examples of five disorders are given. Speech pathology is investigated from these aspects: the…

  1. Hidden Markov models in automatic speech recognition

    NASA Astrophysics Data System (ADS)

    Wrzoskowicz, Adam

    1993-11-01

    This article describes a method for constructing an automatic speech recognition system based on hidden Markov models (HMMs). The author discusses the basic concepts of HMM theory and the application of these models to the analysis and recognition of speech signals. The author provides algorithms which make it possible to train the ASR system and recognize signals on the basis of distinct stochastic models of selected speech sound classes. The author describes the specific components of the system and the procedures used to model and recognize speech. The author discusses problems associated with the choice of optimal signal detection and parameterization characteristics and their effect on the performance of the system. The author presents different options for the choice of speech signal segments and their consequences for the ASR process. The author gives special attention to the use of lexical, syntactic, and semantic information for the purpose of improving the quality and efficiency of the system. The author also describes an ASR system developed by the Speech Acoustics Laboratory of the IBPT PAS. The author discusses the results of experiments on the effect of noise on the performance of the ASR system and describes methods of constructing HMM's designed to operate in a noisy environment. The author also describes a language for human-robot communications which was defined as a complex multilevel network from an HMM model of speech sounds geared towards Polish inflections. The author also added mandatory lexical and syntactic rules to the system for its communications vocabulary.

  2. Multi-input multi-output underwater communications over sparse and frequency modulated acoustic channels.

    PubMed

    Ling, Jun; Zhao, Kexin; Li, Jian; Nordenvaad, Magnus Lundberg

    2011-07-01

    This paper addresses multi-input multi-output (MIMO) communications over sparse acoustic channels suffering from frequency modulations. An extension of the recently introduced SLIM algorithm, which stands for sparse learning via iterative minimization, is presented to estimate the sparse and frequency modulated acoustic channels. The extended algorithm is referred to as generalization of SLIM (GoSLIM). The sparseness is exploited through a hierarchical Bayesian model, and because GoSLIM is user parameter free, it is easy to use in practical applications. Moreover this paper considers channel equalization and symbol detection for various MIMO transmission schemes, including both space-time block coding and spatial multiplexing, under the challenging channel conditions. The effectiveness of the proposed approaches is demonstrated using in-water experimental measurements recently acquired during WHOI09 and ACOMM10 experiments.

  3. Technology for Real-Time Acoustic Communications and Navigation Under Arctic Sea Ice

    NASA Astrophysics Data System (ADS)

    Freitag, L. E.; Ball, K.; Singh, S.; Koski, P.; Partan, J.; Morozov, A.

    2013-12-01

    The use of gliders, floats and powered autonomous underwater vehicles beneath Arctic ice is challenging because surfacing for GPS fixes is risky and also subject to potentially long delays when the ice cover is very dense. For synoptic studies that involve sensors both on the ice and beneath the ice, it is not possible to use fixed transponders on the sea floor, and instead, acoustic sources that are ice-tethered are the best option. However, ice drifts and every transmission is from a different location, and thus the position of the acoustic source must be broadcast as well. We have developed and are preparing to demonstrate a real-time ice-tethered acoustic positioning system that operates at ranges to approximately 100 km using signals at 900 Hz. The system incorporates digital acoustic communication for sending source location and control information, which may be used to re-task the autonomous systems. While the current version is one-way because the mobile platforms are small, larger AUVs (0.3 m dia. or greater) are capable of carrying low-frequency sources and can utilize the system for bidirectional communication. Progress to-date includes a test north of Alaska in 2010 at ranges to 75 km, and in 2011, in the Fram Strait to ranges of 90 km. In both cases data rates at the maximum ranges were low, several bits per second, though at shorter ranges (30-50 km) data rates of 10-40 bps were possible. However, these low data rates are sufficient to transmit 8-12 bytes of location information plus commands to specific units. Next steps in the development and validation of the system include September 2013, again in the Fram Strait, followed by deployment north of Alaska for the ONR Marginal Ice Zone 2014 field campaign.

  4. A Group Intervention Model for Speech and Communication Skills in Patients with Parkinson's Disease: Initial Observations

    ERIC Educational Resources Information Center

    Manor, Yael; Posen, Jennie; Amir, Ofer; Dori, Nechama; Giladi, Nir

    2005-01-01

    Various speech and voice disorders affect 70% to 85% of individuals with Parkinson's disease (PD). Speech treatment is generally conducted on an individual basis, with family member involvement. Clinical experience indicates that many patients do not practice treatments at home or apply the learned techniques in everyday situations. An…

  5. Speech Communication Anxiety: An Impediment to Academic Achievement in the University Classroom.

    ERIC Educational Resources Information Center

    Boohar, Richard K.; Seiler, William J.

    1982-01-01

    The achievement levels of college students taking a bioethics course who demonstrated high and low degrees of speech anxiety were studied. Students with high speech anxiety interacted less with instructors and did not achieve as well as other students. Strategies instructors can use to help students are suggested. (Authors/PP)

  6. Exploring the Speech-Language Connection. Communication and Language Intervention Series, Volume 8.

    ERIC Educational Resources Information Center

    Paul, Rhea, Ed.

    This edited volume investigates the connections between the earliest human sounds and subsequent language development. With special attention to tracheostomies, Downs syndrome, deafness, and speech-motor impairments, the interaction between speech and language is examined in typical development as well as the effect that the interaction has on…

  7. Multilingual Aspects of Speech Sound Disorders in Children. Communication Disorders across Languages

    ERIC Educational Resources Information Center

    McLeod, Sharynne; Goldstein, Brian

    2012-01-01

    Multilingual Aspects of Speech Sound Disorders in Children explores both multilingual and multicultural aspects of children with speech sound disorders. The 30 chapters have been written by 44 authors from 16 different countries about 112 languages and dialects. The book is designed to translate research into clinical practice. It is divided into…

  8. Prompting Students to Contemplate Effective Communication with a Visible Speech Chart from the 1870s

    ERIC Educational Resources Information Center

    Potter, Lee Ann

    2014-01-01

    In this article, director of Educational Outreach at the Library of Congress Lee Ann Potter describes a classroom activity that focuses on an 1876 single-page circular published in Salem, Massachusetts about Alexander Melville Bell's Visible Speech. A. M. Bell's son, Alexander Graham Bell described "Visible Speech" as "a…

  9. Formant-frequency variation and informational masking of speech by extraneous formants: evidence against dynamic and speech-specific acoustical constraints.

    PubMed

    Roberts, Brian; Summers, Robert J; Bailey, Peter J

    2014-08-01

    How speech is separated perceptually from other speech remains poorly understood. Recent research indicates that the ability of an extraneous formant to impair intelligibility depends on the variation of its frequency contour. This study explored the effects of manipulating the depth and pattern of that variation. Three formants (F1+F2+F3) constituting synthetic analogues of natural sentences were distributed across the 2 ears, together with a competitor for F2 (F2C) that listeners must reject to optimize recognition (left = F1+F2C; right = F2+F3). The frequency contours of F1 - F3 were each scaled to 50% of their natural depth, with little effect on intelligibility. Competitors were created either by inverting the frequency contour of F2 about its geometric mean (a plausibly speech-like pattern) or using a regular and arbitrary frequency contour (triangle wave, not plausibly speech-like) matched to the average rate and depth of variation for the inverted F2C. Adding a competitor typically reduced intelligibility; this reduction depended on the depth of F2C variation, being greatest for 100%-depth, intermediate for 50%-depth, and least for 0%-depth (constant) F2Cs. This suggests that competitor impact depends on overall depth of frequency variation, not depth relative to that for the target formants. The absence of tuning (i.e., no minimum in intelligibility for the 50% case) suggests that the ability to reject an extraneous formant does not depend on similarity in the depth of formant-frequency variation. Furthermore, triangle-wave competitors were as effective as their more speech-like counterparts, suggesting that the selection of formants from the ensemble also does not depend on speech-specific constraints. PMID:24842068

  10. Early recognition of speech

    PubMed Central

    Remez, Robert E; Thomas, Emily F

    2013-01-01

    Classic research on the perception of speech sought to identify minimal acoustic correlates of each consonant and vowel. In explaining perception, this view designated momentary components of an acoustic spectrum as cues to the recognition of elementary phonemes. This conceptualization of speech perception is untenable given the findings of phonetic sensitivity to modulation independent of the acoustic and auditory form of the carrier. The empirical key is provided by studies of the perceptual organization of speech, a low-level integrative function that finds and follows the sensory effects of speech amid concurrent events. These projects have shown that the perceptual organization of speech is keyed to modulation; fast; unlearned; nonsymbolic; indifferent to short-term auditory properties; and organization requires attention. The ineluctably multisensory nature of speech perception also imposes conditions that distinguish language among cognitive systems. WIREs Cogn Sci 2013, 4:213–223. doi: 10.1002/wcs.1213 PMID:23926454

  11. Speech masking and cancelling and voice obscuration

    DOEpatents

    Holzrichter, John F.

    2013-09-10

    A non-acoustic sensor is used to measure a user's speech and then broadcasts an obscuring acoustic signal diminishing the user's vocal acoustic output intensity and/or distorting the voice sounds making them unintelligible to persons nearby. The non-acoustic sensor is positioned proximate or contacting a user's neck or head skin tissue for sensing speech production information.

  12. Speech, vocabulary, and the education of children using cochlear implants: oral or total communication?

    PubMed

    Connor, C M; Hieber, S; Arts, H A; Zwolan, T A

    2000-10-01

    This study examines the relationship between the teaching method, oral or total communication, used at children's schools and children's consonant-production accuracy and vocabulary development over time. Children who participated in the study (N = 147) demonstrated profound sensorineural hearing loss and had used cochlear implants for between 6 months and 10 years. Educational programs that used an oral communication (OC) approach focused on the development of spoken language, whereas educational programs that used a total communication (TC) approach focused on the development of language using both signed and spoken language. Using Hierarchical Linear Modeling (HLM) we compared the consonant-production accuracy, receptive spoken vocabulary, and expressive spoken and/or signed vocabulary skills, over time, of children who were enrolled in schools that used either OC or TC approaches, while controlling for a number of variables. These variables included age at implantation, preoperative aided speech detection thresholds, type of cochlear implant device used, and whether a complete or incomplete active electrode array was implanted. The results of this study indicated that as they used their implants the children demonstrated improved consonant-production accuracy and expressive and receptive vocabulary over time, regardless of whether their school employed a TC or OC teaching method. Furthermore, there appeared to be a complex relationship among children's performance with the cochlear implant, age at implantation, and communication/teaching strategy employed by the school. Controlling for all variables, children in OC programs demonstrated, on average, superior consonant-production accuracy, with significantly greater rates of improvement in consonant-production accuracy scores over time compared to children in TC programs. However, there was no significant difference between OC and TC groups in performance or rate of growth in consonant-production accuracy when

  13. Design Foundations for Content-Rich Acoustic Interfaces: Investigating Audemes as Referential Non-Speech Audio Cues

    ERIC Educational Resources Information Center

    Ferati, Mexhid Adem

    2012-01-01

    To access interactive systems, blind and visually impaired users can leverage their auditory senses by using non-speech sounds. The current structure of non-speech sounds, however, is geared toward conveying user interface operations (e.g., opening a file) rather than large theme-based information (e.g., a history passage) and, thus, is ill-suited…

  14. Experimental Study of Doppler Effect for Underwater Acoustic Communication Using Orthogonal Signal Division Multiplexing

    NASA Astrophysics Data System (ADS)

    Ebihara, Tadashi; Mizutani, Keiichi

    2012-07-01

    This paper is about the underwater acoustic (UWA) communication using orthogonal signal division multiplexing (OSDM) in shallow water, whose environment is time spread and frequency spread. In this paper, the Doppler effect - Doppler shift and spread - for UWA communication using OSDM is mainly considered. The effects of Doppler shift and Doppler spread are evaluated in a test tank with a moving platform on a stable water surface and with a stable platform with a moving water surface, respectively. Doppler shift correction, which has been considered in simulation-based studies, is found to work effectively. In relation to the effect of Doppler spread, the experimental result well agrees with the simulation result. Through this study, it is confirmed that a smaller frame length is preferable because it enables the measurement of the UWA channel frequently so that it can keep up with channel changes.

  15. An Acoustic OFDM System with Symbol-by-Symbol Doppler Compensation for Underwater Communication

    PubMed Central

    MinhHai, Tran; Rie, Saotome; Suzuki, Taisaku; Wada, Tomohisa

    2016-01-01

    We propose an acoustic OFDM system for underwater communication, specifically for vertical link communications such as between a robot in the sea bottom and a mother ship in the surface. The main contributions are (1) estimation of time varying Doppler shift using continual pilots in conjunction with monitoring the drift of Power Delay Profile and (2) symbol-by-symbol Doppler compensation in frequency domain by an ICI matrix representing nonuniform Doppler. In addition, we compare our proposal against a resampling method. Simulation and experimental results confirm that our system outperforms the resampling method when the velocity changes roughly over OFDM symbols. Overall, experimental results taken in Shizuoka, Japan, show our system using 16QAM, and 64QAM achieved a data throughput of 7.5 Kbit/sec with a transmitter moving at maximum 2 m/s, in a complicated trajectory, over 30 m vertically. PMID:27057558

  16. A self-organizing neural network architecture for auditory and speech perception with applications to acoustic and other temporal prediction problems

    NASA Astrophysics Data System (ADS)

    Cohen, Michael; Grossberg, Stephen

    1994-09-01

    This project is developing autonomous neural network models for the real-time perception and production of acoustic and speech signals. Our SPINET pitch model was developed to take realtime acoustic input and to simulate the key pitch data. SPINET was embedded into a model for auditory scene analysis, or how the auditory system separates sound sources in environments with multiple sources. The model groups frequency components based on pitch and spatial location cues and resonantly binds them within different streams. The model simulates psychophysical grouping data, such as how an ascending, tone groups with a descending tone even if noise exists at the intersection point, and how a tone before and after a noise burst is perceived to continue through the noise. These resonant streams input to working memories, wherein phonetic percepts adapt to global speech rate. Computer simulations quantitatively generate the experimentally observed category boundary shifts for voiced stop pairs that have the same or different place of articulation, including why the interval to hear a double (geminate) stop is twice as long as that to hear two different stops. This model also uses resonant feedback, here between list categories and working memory.

  17. Seabird acoustic communication at sea: a new perspective using bio-logging devices

    PubMed Central

    Thiebault, Andréa; Pistorius, Pierre; Mullers, Ralf; Tremblay, Yann

    2016-01-01

    Most seabirds are very noisy at their breeding colonies, when aggregated in high densities. Calls are used for individual recognition and also emitted during agonistic interactions. When at sea, many seabirds aggregate over patchily distributed resources and may benefit from foraging in groups. Because these aggregations are so common, it raises the question of whether seabirds use acoustic communication when foraging at sea? We deployed video-cameras with built in microphones on 36 Cape gannets (Morus capensis) during the breeding season of 2010–2011 at Bird Island (Algoa Bay, South Africa) to study their foraging behaviour and vocal activity at sea. Group formation was derived from the camera footage. During ~42 h, calls were recorded on 72 occasions from 16 birds. Vocalization exclusively took place in the presence of conspecifics, and mostly in feeding aggregations (81% of the vocalizations). From the observation of the behaviours of birds associated with the emission of calls, we suggest that the calls were emitted to avoid collisions between birds. Our observations show that at least some seabirds use acoustic communication when foraging at sea. These findings open up new perspectives for research on seabirds foraging ecology and their interactions at sea. PMID:27492779

  18. Seabird acoustic communication at sea: a new perspective using bio-logging devices.

    PubMed

    Thiebault, Andréa; Pistorius, Pierre; Mullers, Ralf; Tremblay, Yann

    2016-01-01

    Most seabirds are very noisy at their breeding colonies, when aggregated in high densities. Calls are used for individual recognition and also emitted during agonistic interactions. When at sea, many seabirds aggregate over patchily distributed resources and may benefit from foraging in groups. Because these aggregations are so common, it raises the question of whether seabirds use acoustic communication when foraging at sea? We deployed video-cameras with built in microphones on 36 Cape gannets (Morus capensis) during the breeding season of 2010-2011 at Bird Island (Algoa Bay, South Africa) to study their foraging behaviour and vocal activity at sea. Group formation was derived from the camera footage. During ~42 h, calls were recorded on 72 occasions from 16 birds. Vocalization exclusively took place in the presence of conspecifics, and mostly in feeding aggregations (81% of the vocalizations). From the observation of the behaviours of birds associated with the emission of calls, we suggest that the calls were emitted to avoid collisions between birds. Our observations show that at least some seabirds use acoustic communication when foraging at sea. These findings open up new perspectives for research on seabirds foraging ecology and their interactions at sea. PMID:27492779

  19. Frequency-selective fading statistics of shallow-water acoustic communication channel with a few multipaths

    NASA Astrophysics Data System (ADS)

    Bae, Minja; Park, Jihyun; Kim, Jongju; Xue, Dandan; Park, Kyu-Chil; Yoon, Jong Rak

    2016-07-01

    The bit error rate of an underwater acoustic communication system is related to multipath fading statistics, which determine the signal-to-noise ratio. The amplitude and delay of each path depend on sea surface roughness, propagation medium properties, and source-to-receiver range as a function of frequency. Therefore, received signals will show frequency-dependent fading. A shallow-water acoustic communication channel generally shows a few strong multipaths that interfere with each other and the resulting interference affects the fading statistics model. In this study, frequency-selective fading statistics are modeled on the basis of the phasor representation of the complex path amplitude. The fading statistics distribution is parameterized by the frequency-dependent constructive or destructive interference of multipaths. At a 16 m depth with a muddy bottom, a wave height of 0.2 m, and source-to-receiver ranges of 100 and 400 m, fading statistics tend to show a Rayleigh distribution at a destructive interference frequency, but a Rice distribution at a constructive interference frequency. The theoretical fading statistics well matched the experimental ones.

  20. Ant pupae employ acoustics to communicate social status in their colony's hierarchy.

    PubMed

    Casacci, Luca P; Thomas, Jeremy A; Sala, Marco; Treanor, David; Bonelli, Simona; Balletto, Emilio; Schönrogge, Karsten

    2013-02-18

    The possession of an efficient communication system and an ability to distinguish between young stages are essential attributes that enable eusocial insects to live in complex integrated societies. Although ants communicate primarily via chemicals, it is increasingly clear that acoustical signals also convey important information, including status, between adults in many species. However, all immature stages were believed to be mute. We confirm that larvae and recently formed pupae of Myrmica ants are mute, yet once they are sclerotized, the pupae possess a fully functioning stridulatory organ. The sounds generated by worker pupae were similar to those of workers but were emitted as single pulses rather than in the long sequences characteristic of adults; both induced the same range and intensity of benevolent behaviors when played back to unstressed workers. Both white and sclerotized pupae have a higher social status than larvae within Myrmica colonies, but the latter's status fell significantly after they were made mute. Our results suggest that acoustical signals supplant semiochemicals as a means of identification in sclerotized pupae, perhaps because their hardened integuments block the secretion of brood pheromones or because their developing adult secretions initially differ from overall colony odors. PMID:23394832

  1. Effect of temperature on acoustic communication: sound production in the croaking gourami (labyrinth fishes).

    PubMed

    Ladich, Friedrich; Schleinzer, Günter

    2015-04-01

    Sound communication comprising the production and detection of acoustic signals is affected by ambient temperature in ectothermic animals. In the present study we investigated the effects of temperature on sound production and characteristics in the croaking gourami Trichopsis vittata, a freshwater fish from Southeast Asia possessing a highly specialized sound-generating mechanism found only in a single genus. The croaking gourami produces pulsed sounds by stretching and plucking two enhanced pectoral fin tendons during rapid pectoral fin beating. Croaking sounds typically consist of a series of double-pulsed bursts with main energies between 1 and 1.5 kHz. Sounds were recorded during dyadic contests between two males at three different temperatures (25°, 30° and 35°C). The mean dominant frequency increased with rising temperature from 1.18 to 1.33 kHz, whereas temporal characteristics decreased. The sound interval dropped from 492 to 259 ms, the burst period from 51 to 35 ms and the pulse period from 5.8 to 5.1 ms. In contrast, the number of sounds and number of bursts within a sound were not affected by temperature. The current study shows that spectral and temporal characteristics of sounds are affected in different ways by temperature in the croaking gourami, whereas the numbers of sounds and bursts remain unaffected. We conclude that acoustic communication in gouramis is affected by changes in ambient temperature. PMID:25433336

  2. Language, Speech, and Communication Skills Training: Abstracts of Doctoral Dissertations Published in "Dissertation Abstracts International," January through June 1978 (Vol. 38 Nos. 7 through 12).

    ERIC Educational Resources Information Center

    ERIC Clearinghouse on Reading and Communication Skills, Urbana, IL.

    This collection of abstracts is part of a continuing series providing information on recent doctoral dissertations. The 21 titles deal with a variety of topics, including the following: communications curricula for technical education programs; the recorded oral communication of business students; interactions between parents and speech-language…

  3. A Review of Preservice Training in Augmentative and Alternative Communication for Speech-Language Pathologists, Special Education Teachers, and Occupational Therapists

    ERIC Educational Resources Information Center

    Costigan, F. Aileen; Light, Janice

    2010-01-01

    Speech-language pathologists, special education teachers, and occupational therapists are all likely to encounter individuals with complex communication needs who require augmentative and alternative communication (AAC) in clinical and educational practice. The research on preservice AAC training for these professionals was thus reviewed to…

  4. A Graphic Symbol Tool for the Evaluation of Communication, Satisfaction and Priorities of Individuals with Intellectual Disability Who Use a Speech Generating Device

    ERIC Educational Resources Information Center

    Valiquette, Christine; Sutton, Ann; Ska, Bernadette

    2010-01-01

    This article reports on the views of individuals with learning disability (LD) on their use of their speech generating devices (SGDs), their satisfaction about their communication, and their priorities. The development of an interview tool made of graphic symbols and entitled Communication, Satisfaction and Priorities of SGD Users (CSPU) is…

  5. Speech articulator measurements using low power EM-wave sensors

    SciTech Connect

    Holzrichter, J.F.; Burnett, G.C.; Ng, L.C.; Lea, W.A.

    1998-01-01

    Very low power electromagnetic (EM) wave sensors are being used to measure speech articulator motions as speech is produced. Glottal tissue oscillations, jaw, tongue, soft palate, and other organs have been measured. Previously, microwave imaging (e.g., using radar sensors) appears not to have been considered for such monitoring. Glottal tissue movements detected by radar sensors correlate well with those obtained by established laboratory techniques, and have been used to estimate a voiced excitation function for speech processing applications. The noninvasive access, coupled with the small size, low power, and high resolution of these new sensors, permit promising research and development applications in speech production, communication disorders, speech recognition and related topics. {copyright} {ital 1998 Acoustical Society of America.}

  6. Production and perception of listener-oriented clear speech in child language.

    PubMed

    Syrett, Kristen; Kawahara, Shigeto

    2014-11-01

    In this paper, we ask whether children are sensitive to the needs of their interlocutor, and, if so, whether they - like adults - modify acoustic characteristics of their speech as part of a communicative goal. In a production task, preschoolers participated in a word learning task that favored the use of clear speech. Children produced vowels that were longer, more intense, more dispersed in the vowel space, and had a more expanded F0 range than normal speech. Two perception studies with adults showed that these acoustic differences were perceptible and were used to distinguish normal and clear speech styles. We conclude that preschoolers are sensitive to aspects of the speaker-hearer relationship calling upon them to modify their speech in ways that benefit their listener.

  7. Psychoacoustic cues to emotion in speech prosody and music.

    PubMed

    Coutinho, Eduardo; Dibben, Nicola

    2013-01-01

    There is strong evidence of shared acoustic profiles common to the expression of emotions in music and speech, yet relatively limited understanding of the specific psychoacoustic features involved. This study combined a controlled experiment and computational modelling to investigate the perceptual codes associated with the expression of emotion in the acoustic domain. The empirical stage of the study provided continuous human ratings of emotions perceived in excerpts of film music and natural speech samples. The computational stage created a computer model that retrieves the relevant information from the acoustic stimuli and makes predictions about the emotional expressiveness of speech and music close to the responses of human subjects. We show that a significant part of the listeners' second-by-second reported emotions to music and speech prosody can be predicted from a set of seven psychoacoustic features: loudness, tempo/speech rate, melody/prosody contour, spectral centroid, spectral flux, sharpness, and roughness. The implications of these results are discussed in the context of cross-modal similarities in the communication of emotion in the acoustic domain. PMID:23057507

  8. Psychoacoustic cues to emotion in speech prosody and music.

    PubMed

    Coutinho, Eduardo; Dibben, Nicola

    2013-01-01

    There is strong evidence of shared acoustic profiles common to the expression of emotions in music and speech, yet relatively limited understanding of the specific psychoacoustic features involved. This study combined a controlled experiment and computational modelling to investigate the perceptual codes associated with the expression of emotion in the acoustic domain. The empirical stage of the study provided continuous human ratings of emotions perceived in excerpts of film music and natural speech samples. The computational stage created a computer model that retrieves the relevant information from the acoustic stimuli and makes predictions about the emotional expressiveness of speech and music close to the responses of human subjects. We show that a significant part of the listeners' second-by-second reported emotions to music and speech prosody can be predicted from a set of seven psychoacoustic features: loudness, tempo/speech rate, melody/prosody contour, spectral centroid, spectral flux, sharpness, and roughness. The implications of these results are discussed in the context of cross-modal similarities in the communication of emotion in the acoustic domain.

  9. Experimental investigation of the effects of the acoustical conditions in a simulated classroom on speech recognition and learning in children a

    PubMed Central

    Valente, Daniel L.; Plevinsky, Hallie M.; Franco, John M.; Heinrichs-Graham, Elizabeth C.; Lewis, Dawna E.

    2012-01-01

    The potential effects of acoustical environment on speech understanding are especially important as children enter school where students’ ability to hear and understand complex verbal information is critical to learning. However, this ability is compromised because of widely varied and unfavorable classroom acoustics. The extent to which unfavorable classroom acoustics affect children’s performance on longer learning tasks is largely unknown as most research has focused on testing children using words, syllables, or sentences as stimuli. In the current study, a simulated classroom environment was used to measure comprehension performance of two classroom learning activities: a discussion and lecture. Comprehension performance was measured for groups of elementary-aged students in one of four environments with varied reverberation times and background noise levels. The reverberation time was either 0.6 or 1.5 s, and the signal-to-noise level was either +10 or +7 dB. Performance is compared to adult subjects as well as to sentence-recognition in the same condition. Significant differences were seen in comprehension scores as a function of age and condition; both increasing background noise and reverberation degraded performance in comprehension tasks compared to minimal differences in measures of sentence-recognition. PMID:22280587

  10. Acoustic communication in the Greater Sage-Grouse (Centrocercus urophasianus) an examination into vocal sacs, sound propagation, and signal directionality

    NASA Astrophysics Data System (ADS)

    Dantzker, Marc Steven

    The thesis is an inquiry into the acoustic communication of a very unusual avian species, the Greater Sage-Grouse, Centrocercus urophasianus. One of the most outstanding features of this animal's dynamic mating display is its use of paired air sacs that emerge explosively from an esophageal pouch. My first line of inquiry into this system is a review of the form and function of similar vocal apparatuses, collectively called vocal sacs, in birds. Next, with a combination of mathematical models and field measurements, My collaborator and I investigate the acoustic environment where the Greater Sage-Grouse display. The complexities of this acoustic environment are relevant both to the birds and to the subsequent examinations of the display's properties. Finally, my collaborators and I examine a cryptic component of the acoustic display --- directionality --- which we measured simultaneously from multiple locations around free moving grouse on their mating grounds.

  11. Towards a Sign Language Synthesizer: a Bridge to Communication Gap of the Hearing/Speech Impaired Community

    NASA Astrophysics Data System (ADS)

    Maarif, H. A.; Akmeliawati, R.; Gunawan, T. S.; Shafie, A. A.

    2013-12-01

    Sign language synthesizer is a method to visualize the sign language movement from the spoken language. The sign language (SL) is one of means used by HSI people to communicate to normal people. But, unfortunately the number of people, including the HSI people, who are familiar with sign language is very limited. These cause difficulties in the communication between the normal people and the HSI people. The sign language is not only hand movement but also the face expression. Those two elements have complimentary aspect each other. The hand movement will show the meaning of each signing and the face expression will show the emotion of a person. Generally, Sign language synthesizer will recognize the spoken language by using speech recognition, the grammatical process will involve context free grammar, and 3D synthesizer will take part by involving recorded avatar. This paper will analyze and compare the existing techniques of developing a sign language synthesizer, which leads to IIUM Sign Language Synthesizer.

  12. The Swedish Speech Interpretation Service: an exploratory study of a new communication support provided to people with aphasia.

    PubMed

    Larsson, Inger; Thorén-Jönsson, Anna-Lisa

    2007-12-01

    Until recently, people with complex communication needs have had no access to professional interpreters. The Swedish Speech Interpretation Service (SSIS) is attempting to address this problem. This qualitative study reports on how 12 persons with aphasia experienced the services of a professional interpreter from the SSIS. The results are presented in two themes: (a) The purpose of using an interpreter, which addresses issues relating to autonomy, privacy, and burden on family members; and (b) perceptions of quality of service, which addresses issues relating to the skills and professionalism of and accessibility to interpreters. Results highlight the ongoing need for the SSIS and its importance to the participants. The professional interpreter as an augmentative and alternative communication (AAC) "tool" that could be used to enhance the participation of people with aphasia in the community-at-large, is also discussed; as is whether relatives and other non-professionals or professional interpreters should interpret for people with aphasia.

  13. The Swedish Speech Interpretation Service: an exploratory study of a new communication support provided to people with aphasia.

    PubMed

    Larsson, Inger; Thorén-Jönsson, Anna-Lisa

    2007-01-01

    Until recently, people with complex communication needs have had no access to professional interpreters. The Swedish Speech Interpretation Service (SSIS) is attempting to address this problem. This qualitative study reports on how 12 persons with aphasia experienced the services of a professional interpreter from the SSIS. The results are presented in two themes: (a) The purpose of using an interpreter, which addresses issues relating to autonomy, privacy, and burden on family members; and (b) perceptions of quality of service, which addresses issues relating to the skills and professionalism of and accessibility to interpreters. Results highlight the ongoing need for the SSIS and its importance to the participants. The professional interpreter as an augmentative and alternative communication (AAC) "tool" that could be used to enhance the participation of people with aphasia in the community-at large, is also discussed; as is whether relatives and other non-professionals or professional interpreters should interpret for people with aphasia.

  14. Tracking the sparseness of the underlying support in shallow water acoustic communications

    NASA Astrophysics Data System (ADS)

    Sen Gupta, Ananya; Preisig, James

    2012-06-01

    Tracking the shallow water acoustic channel in real time poses an open challenge towards improving the data rate in high-speed underwater communications. Multipath arrivals due to reflection from the moving ocean surface and the sea bottom, along with surface wave focusing events, lead to a rapidly fluctuating complex-valued channel impulse response and associated Delay-Doppler spread function that follow heavy-tailed distributions. The sparse channel or Delay-Doppler spread function components are difficult to track in real time using popular sparse sensing techniques due to the coherent and dynamic nature of the optimization problem as well as the timevarying and potentially non-stationary sparseness of the underlying support. We build on related work using non-convex optimization to track the shallow water acoustic channel in real time at high precision and tracking speed to develop strategies to estimate the non-stationary sparseness of the underlying support. Specifically, we employ non-convex manifold navigational techniques to estimate the support sparseness to balance the weighting between the L1 norm of the tracked coefficients and the L2 norm of the estimation error. We explore the efficacy of our methods against experimental field data collected at 200 meters range, 15 meters depth and varying wind conditions.

  15. Design of a robust underwater acoustic communication system over multipath fading channels

    NASA Astrophysics Data System (ADS)

    Kim, Jangeun; Shim, Taebo

    2012-11-01

    Due to the surface and bottom constraints of the underwater acoustic channel (UAC) in shallow waters, multipath fading occurs and causes degradation of the signal for the UAC system. To overcome these problems, a robust underwater acoustic communication system was designed over multipath fading channels by employing both decision feedback equalization with the RLS algorithm and convolutional coding with interleaving+shuffling block data sequence. The dual use of these two methods simultaneously can reduce the intersymbol interference (ISI) and the adjacent bit and burst errors. The system will retransmit the same signal if the system fails to estimate the channel due to severe multipath fading. To verify the performance of the system, the transmission of an image was tested using a 524,288bit gray-scaled image through the multipath fading channel. The test results showed that the number of bit errors was reduced from 86,824 to 5,106 when the reference SNR was 12 dB.

  16. The involvement of people with lifelong disability and communication impairment in lecturing to speech-language pathology students.

    PubMed

    Balandin, Susan; Hines, Monique

    2011-10-01

    In order to provide appropriate service and support to people with lifelong disability, including those who use augmentative and alternative communication (AAC), speech-language pathologists need to develop positive attitudes to people with disability and a holistic approach. The aim of this paper was to explore final year students' and new graduates' perceptions of the impact of lectures by people with lifelong disability on their attitudes and learning. Fourteen final year speech-language pathology students and nine graduates who had been practising for a year participated in four focus groups to discuss their perceptions of four 50-minute lectures by people with lifelong disability and communication impairment. A content analysis of the focus group transcripts revealed five themes; understanding the perspectives of people with disability, seeing the person as a whole, attitudes, working in the field, and gaps in the lectures. Overall there were few differences between the groups that were not accounted for by clinical experience. Participants agreed that the lectures were interesting and informative and provided new insights into lifelong disability, but were an adjunct to a learning process that also required theoretical lectures or clinical practice.

  17. Differential Diagnosis of Severe Speech Disorders Using Speech Gestures

    ERIC Educational Resources Information Center

    Bahr, Ruth Huntley

    2005-01-01

    The differentiation of childhood apraxia of speech from severe phonological disorder is a common clinical problem. This article reports on an attempt to describe speech errors in children with childhood apraxia of speech on the basis of gesture use and acoustic analyses of articulatory gestures. The focus was on the movement of articulators and…

  18. Speech coding

    SciTech Connect

    Ravishankar, C., Hughes Network Systems, Germantown, MD

    1998-05-08

    Speech is the predominant means of communication between human beings and since the invention of the telephone by Alexander Graham Bell in 1876, speech services have remained to be the core service in almost all telecommunication systems. Original analog methods of telephony had the disadvantage of speech signal getting corrupted by noise, cross-talk and distortion Long haul transmissions which use repeaters to compensate for the loss in signal strength on transmission links also increase the associated noise and distortion. On the other hand digital transmission is relatively immune to noise, cross-talk and distortion primarily because of the capability to faithfully regenerate digital signal at each repeater purely based on a binary decision. Hence end-to-end performance of the digital link essentially becomes independent of the length and operating frequency bands of the link Hence from a transmission point of view digital transmission has been the preferred approach due to its higher immunity to noise. The need to carry digital speech became extremely important from a service provision point of view as well. Modem requirements have introduced the need for robust, flexible and secure services that can carry a multitude of signal types (such as voice, data and video) without a fundamental change in infrastructure. Such a requirement could not have been easily met without the advent of digital transmission systems, thereby requiring speech to be coded digitally. The term Speech Coding is often referred to techniques that represent or code speech signals either directly as a waveform or as a set of parameters by analyzing the speech signal. In either case, the codes are transmitted to the distant end where speech is reconstructed or synthesized using the received set of codes. A more generic term that is applicable to these techniques that is often interchangeably used with speech coding is the term voice coding. This term is more generic in the sense that the

  19. Exploring School Life from the Lens of a Child Who Does Not Use Speech to Communicate

    ERIC Educational Resources Information Center

    Ajodhia-Andrews, Amanda; Berman, Rachel

    2009-01-01

    The "new sociology of childhood" emphasizes listening to the voices of children when conducting research about their lives. In keeping with this framework, the following case study highlights the use of inclusive strategies and the importance of the researcher's orientation in exploring the perspectives of a child who does not use speech to…

  20. Important and Unimportant Organizational Communication: Public Employee Freedom of Speech after "Connick v. Myers."

    ERIC Educational Resources Information Center

    Sanders, Wayne C.

    A review of 16 Federal Court of Appeals cases indicates the impact of the "Connick v. Myers" case on the nature of freedom of speech in public organizations. The case involved the firing of an assistant district attorney for circulating a job satisfaction survey after she was transferred to a less desirable section of the courts. Since the 1983…

  1. Minke whale song, spacing, and acoustic communication on the Great Barrier Reef, Australia

    NASA Astrophysics Data System (ADS)

    Gedamke, Jason

    associated contextual data of recorded sounds were analyzed. Two categories of sound are described here: (1) patterned song, which was regularly repeated in one of three patterns: slow, fast, and rapid-clustered repetition, and (2) non-patterned "social" sounds recorded from gregarious assemblages of whales. These discrete acoustic signals may comprise a graded system of communication (Slow/fast song → Rapid-clustered song → Social sounds) that is related to the spacing between whales.

  2. Maternal depression and the learning-promoting effects of infant-directed speech: Roles of maternal sensitivity, depression diagnosis, and speech acoustic cues.

    PubMed

    Kaplan, Peter S; Danko, Christina M; Cejka, Anna M; Everhart, Kevin D

    2015-11-01

    The hypothesis that the associative learning-promoting effects of infant-directed speech (IDS) depend on infants' social experience was tested in a conditioned-attention paradigm with a cumulative sample of 4- to 14-month-old infants. Following six forward pairings of a brief IDS segment and a photographic slide of a smiling female face, infants of clinically depressed mothers exhibited evidence of having acquired significantly weaker voice-face associations than infants of non-depressed mothers. Regression analyses revealed that maternal depression was significantly related to infant learning even after demographic correlates of depression, antidepressant medication use, and extent of pitch modulation in maternal IDS had been taken into account. However, after maternal depression had been accounted for, maternal emotional availability, coded by blind raters from separate play interactions, accounted for significant further increments in the proportion of variance accounted for in infant learning scores. Both maternal depression and maternal insensitivity negatively, and additively, predicted poor learning.

  3. Study of environmental sound source identification based on hidden Markov model for robust speech recognition

    NASA Astrophysics Data System (ADS)

    Nishiura, Takanobu; Nakamura, Satoshi

    2003-10-01

    Humans communicate with each other through speech by focusing on the target speech among environmental sounds in real acoustic environments. We can easily identify the target sound from other environmental sounds. For hands-free speech recognition, the identification of the target speech from environmental sounds is imperative. This mechanism may also be important for a self-moving robot to sense the acoustic environments and communicate with humans. Therefore, this paper first proposes hidden Markov model (HMM)-based environmental sound source identification. Environmental sounds are modeled by three states of HMMs and evaluated using 92 kinds of environmental sounds. The identification accuracy was 95.4%. This paper also proposes a new HMM composition method that composes speech HMMs and an HMM of categorized environmental sounds for robust environmental sound-added speech recognition. As a result of the evaluation experiments, we confirmed that the proposed HMM composition outperforms the conventional HMM composition with speech HMMs and a noise (environmental sound) HMM trained using noise periods prior to the target speech in a captured signal. [Work supported by Ministry of Public Management, Home Affairs, Posts and Telecommunications of Japan.

  4. Effect of reflected and refracted signals on coherent underwater acoustic communication: results from the Kauai experiment (KauaiEx 2003).

    PubMed

    Rouseff, Daniel; Badiey, Mohsen; Song, Aijun

    2009-11-01

    The performance of a communications equalizer is quantified in terms of the number of acoustic paths that are treated as usable signal. The analysis uses acoustical and oceanographic data collected off the Hawaiian Island of Kauai. Communication signals were measured on an eight-element vertical array at two different ranges, 1 and 2 km, and processed using an equalizer based on passive time-reversal signal processing. By estimating the Rayleigh parameter, it is shown that all paths reflected by the sea surface at both ranges undergo incoherent scattering. It is demonstrated that some of these incoherently scattered paths are still useful for coherent communications. At range of 1 km, optimal communications performance is achieved when six acoustic paths are retained and all paths with more than one reflection off the sea surface are rejected. Consistent with a model that ignores loss from near-surface bubbles, the performance improves by approximately 1.8 dB when increasing the number of retained paths from four to six. The four-path results though are more stable and require less frequent channel estimation. At range of 2 km, ray refraction is observed and communications performance is optimal when some paths with two sea-surface reflections are retained.

  5. Acoustic communication in two freshwater gobies: ambient noise and short-range propagation in shallow streams.

    PubMed

    Lugli, M; Fine, M L

    2003-07-01

    Noise is an important theoretical constraint on the evolution of signal form and sensory performance. In order to determine environmental constraints on the communication of two freshwater gobies Padogobius martensii and Gobius nigricans, numerous noise spectra were measured from quiet areas and ones adjacent to waterfalls and rapids in two shallow stony streams. Propagation of goby sounds and waterfall noise was also measured. A quiet window around 100 Hz is present in many noise spectra from noisy locations. The window lies between two noise sources, a low-frequency one attributed to turbulence, and a high-frequency one (200-500 Hz) attributed to bubble noise from water breaking the surface. Ambient noise from a waterfall (frequencies below 1 kHz) attenuates as much as 30 dB between 1 and 2 m, after which values are variable without further attenuation (i.e., buried in the noise floor). Similarly, courtship sounds of P. martensii attenuate as much as 30 dB between 5 and 50 cm. Since gobies are known to court in noisy as well as quiet locations in these streams, their acoustic communication system (sounds and auditory system) must be able to cope with short-range propagation dictated by shallow depths and ambient noise in noisy locations. PMID:12880062

  6. Acoustic communication in two freshwater gobies: ambient noise and short-range propagation in shallow streams.

    PubMed

    Lugli, M; Fine, M L

    2003-07-01

    Noise is an important theoretical constraint on the evolution of signal form and sensory performance. In order to determine environmental constraints on the communication of two freshwater gobies Padogobius martensii and Gobius nigricans, numerous noise spectra were measured from quiet areas and ones adjacent to waterfalls and rapids in two shallow stony streams. Propagation of goby sounds and waterfall noise was also measured. A quiet window around 100 Hz is present in many noise spectra from noisy locations. The window lies between two noise sources, a low-frequency one attributed to turbulence, and a high-frequency one (200-500 Hz) attributed to bubble noise from water breaking the surface. Ambient noise from a waterfall (frequencies below 1 kHz) attenuates as much as 30 dB between 1 and 2 m, after which values are variable without further attenuation (i.e., buried in the noise floor). Similarly, courtship sounds of P. martensii attenuate as much as 30 dB between 5 and 50 cm. Since gobies are known to court in noisy as well as quiet locations in these streams, their acoustic communication system (sounds and auditory system) must be able to cope with short-range propagation dictated by shallow depths and ambient noise in noisy locations.

  7. Interference suppression for code-division multiple-access communications in an underwater acoustic channel.

    PubMed

    Yang, T C; Yang, Wen-Bin

    2009-07-01

    In a code-division multiple-access communications network, the signal from a nearby user often creates a strong interference for the signal from a distant user. This is known as the near-far problem. Power control of source levels is ineffective in an underwater acoustic channel due to the slow sound speed. Interference rejection based on code orthogonality is ineffective using matched-filter processing due to the fact that multipath arrivals effectively destroy the code orthogonality and that the signal arrival times between different users are not synchronized. An algorithm, called hyperspace cancellation by coordinate zeroing, is used in this paper to remove/suppress interference. Using a fast Walsh-Hadamard transform (FWHT) based on the interferer's code sequence, the interference signal is enhanced and removed by coordinate zeroing. The residual signal is transformed back using an inverse FWHT. The filtered data, with the interference signal largely removed, are processed using the desired signal code sequence. Two methods previously developed for direct-sequence spread-spectrum communications in an underwater channel are used to extract the transmitted symbols. Low bit error rate (<10(-2)) is found with the at-sea data for signal-to-interference ratio as low as -8 to -11 dB.

  8. Improvement of Power Efficiency for Underwater Acoustic Communication Using Orthogonal Signal Division Multiplexing over Multiple Transducers

    NASA Astrophysics Data System (ADS)

    Ebihara, Tadashi

    2013-07-01

    In underwater acoustic (UWA) communication, power efficiency is one of the important characteristics. This paper is about multistream transmission using orthogonal signal division multiplexing (OSDM) as a technique to increase power efficiency. In this work, the performance of multistream transmission using OSDM is evaluated both experimentally in a test tank and by numerical simulation. Through this study, it is confirmed that the multistream transmission scheme is effective in enhancing the power efficiency compared with the single-stream transmission using higher order modulation. Moreover, the performance of multistream transmission using OSDM is compared with the existing scheme, multistream transmission using orthogonal frequency division multiplexing (OFDM). The obtained results suggest that multistream transmission using OSDM is attractive because it can achieve the same bit-error rate (BER) and the same data rate with less power of the signal, compared with the reference. Although the calculation cost of OSDM in the receiver remains as an issue, multistream transmission using OSDM may contribute to high-speed UWA communication because of its excellent power efficiency.

  9. Sweep-spread carrier for underwater communication over acoustic channels with strong multipath propagation.

    PubMed

    Kebkal, Konstantin G; Bannasch, Rudolf

    2002-11-01

    Multipath propagation causes the transmitted signal to take numerous, time-varying different-length paths to the receiver. Exploitation of conventional frequency-constant carrier signals for communication over underwater acoustic channels typically shows that intricate mutual time variations of multipath arrivals create an extremely hard problem for equalizers to compensate for nonstable intersymbol interactions. Communication over such channels can be, however, substantially improved by using a new method based on the implementation of a sweep-spread carrier (S2C). Such a carrier consists of a succession of sweeps, which cause permanent rapid fluctuation of signal frequency. Apart from some additional useful effects such as providing multiuser access and reducing influence of narrow-band noise, the method provides significant processing improvement enabling clear separation of multipath arrivals by converting their time delays into their frequency reallocations--the steeper the sweeps, the better the multipath resolution that can be achieved. When converting the signal into constant intermediate frequencies, the best suitable multipath arrival can be separated not as traditionally in time domains, applying complex equalizers, but in frequency domains by means of usual band-pass filters. High transmission stability of this approach was confirmed in both computer simulations as well as in validation experiments carried out since summer 1999.

  10. Green symphonies: a call for studies on acoustic communication in plants

    PubMed Central

    2013-01-01

    Sound and its use in communication have significantly contributed to shaping the ecology, evolution, behavior, and ultimately the success of many animal species. Yet, the ability to use sound is not a prerogative of animals. Plants may also use sound, but we have been unable to effectively research what the ecological and evolutionary implications might be in a plant’s life. Why should plants emit and receive sound and is there information contained in those sounds? I hypothesize that it would be particularly advantageous for plants to learn about the surrounding environment using sound, as acoustic signals propagate rapidly and with minimal energetic or fitness costs. In fact, both emission and detection of sound may have adaptive value in plants by affecting responses in other organisms, plants, and animals alike. The systematic exploration of the functional, ecological, and evolutionary significance of sound in the life of plants is expected to prompt a reinterpretation of our understanding of these organisms and galvanize the emergence of novel concepts and perspectives on their communicative complexity. PMID:23754865

  11. Perceptions of Communication before and after a Speech Pathology Intervention for an Adult with Intellectual Disability

    ERIC Educational Resources Information Center

    Forster, Sheridan; Iacono, Teresa

    2007-01-01

    Background: The aim of this study was to explore, from the perspective of key communication partners, the meaning of a communication intervention and associated changes for an adult with severe intellectual disability (ID). Method: The parents and key support workers of Mark, a 22-year-old man with ID and complex communication needs, were…

  12. Rate-constrained source separation for speech enhancement in wireless-communicated binaural hearing aids

    NASA Astrophysics Data System (ADS)

    Ayllón, David; Gil-Pita, Roberto; Rosa-Zurera, Manuel

    2013-12-01

    A recent trend in hearing aids is the connection of the left and right devices to collaborate between them. Binaural systems can provide natural binaural hearing and support the improvement of speech intelligibility in noise, but they require data transmission between both devices, which increases the power consumption. This paper presents a novel sound source separation algorithm for binaural speech enhancement based on supervised machine learning and time-frequency masking. The system is designed considering the power restrictions in hearing aids, constraining both the computational cost of the algorithm and the transmission bit rate. The transmission schema is optimized using a tailored evolutionary algorithm that assigns a different number of bits to each frequency band. The proposed algorithm requires less than 10% of the available computational resources for signal processing and obtains good separation performance using bit rates lower than 64 kbps.

  13. Room Acoustics

    NASA Astrophysics Data System (ADS)

    Kuttruff, Heinrich; Mommertz, Eckard

    The traditional task of room acoustics is to create or formulate conditions which ensure the best possible propagation of sound in a room from a sound source to a listener. Thus, objects of room acoustics are in particular assembly halls of all kinds, such as auditoria and lecture halls, conference rooms, theaters, concert halls or churches. Already at this point, it has to be pointed out that these conditions essentially depend on the question if speech or music should be transmitted; in the first case, the criterion for transmission quality is good speech intelligibility, in the other case, however, the success of room-acoustical efforts depends on other factors that cannot be quantified that easily, not least it also depends on the hearing habits of the listeners. In any case, absolutely "good acoustics" of a room do not exist.

  14. Speech Intelligibility

    NASA Astrophysics Data System (ADS)

    Brand, Thomas

    Speech intelligibility (SI) is important for different fields of research, engineering and diagnostics in order to quantify very different phenomena like the quality of recordings, communication and playback devices, the reverberation of auditoria, characteristics of hearing impairment, benefit using hearing aids or combinations of these things.

  15. ON THE NATURE OF SPEECH SCIENCE.

    ERIC Educational Resources Information Center

    PETERSON, GORDON E.

    IN THIS ARTICLE THE NATURE OF THE DISCIPLINE OF SPEECH SCIENCE IS CONSIDERED AND THE VARIOUS BASIC AND APPLIED AREAS OF THE DISCIPLINE ARE DISCUSSED. THE BASIC AREAS ENCOMPASS THE VARIOUS PROCESSES OF THE PHYSIOLOGY OF SPEECH PRODUCTION, THE ACOUSTICAL CHARACTERISTICS OF SPEECH, INCLUDING THE SPEECH WAVE TYPES AND THE INFORMATION-BEARING ACOUSTIC…

  16. Speech, Language, and Communication Disorders. Papers Presented at the Annual International Convention of the Council for Exceptional Children (48th, Chicago, Illinois, April 19-25, 1970).

    ERIC Educational Resources Information Center

    Council for Exceptional Children, Arlington, VA.

    Speech, language, and communication disorders were among the topics of papers presented at the convention of the Council for Exceptional Children in Chicago, 1970. Discussions include evaluation theory and caseload selection by Lear Ashmore, differential diagnosis in a rural school by William L. Shinder, the practical application of differential…

  17. "Off the Radar:" The Framing of Speech, Language and Communication in the Description of Children with Special Educational Needs in Literacy

    ERIC Educational Resources Information Center

    Mroz, Maria A.

    2014-01-01

    This paper considers how pre-service primary teachers in England (trainees) are supported to work with children with special educational needs (SEN) using a personalized learning task. The focus, in particular, considers how speech, language and communication skills are foregrounded in the reports describing children with moderate learning…

  18. Communicative Intentions of Child-Directed Speech in Three Different Learning Environments: Observations from the Netherlands, and Rural and Urban Mozambique

    ERIC Educational Resources Information Center

    Vogt, Paul; Mastin, J. Douglas; Schots, Diede M. A.

    2015-01-01

    This article compares the communicative intentions observed in the speech addressed to children of 1;1 and 1;6 years old from three cultural communities: the Netherlands, rural Mozambique, and urban Mozambique. These communities represent two prototypical learning environments and a third hybrid: Western, urban, middle-class families; non-Western,…

  19. Cost-Effectiveness of Interventions for Children with Speech, Language and Communication Needs (SLCN): A Review Using the Drummond and Jefferson (1996) "Referee's Checklist"

    ERIC Educational Resources Information Center

    Law, James; Zeng, Biao; Lindsay, Geoff; Beecham, Jennifer

    2012-01-01

    Background: Although economic evaluation has been widely recognized as a key feature of both health services and educational research, for many years there has been a paucity of such studies relevant to services for children with speech, language and communication needs (SLCN), making the application of economic arguments to the development of…

  20. The Impact of the Picture Exchange Communication System on Requesting and Speech Development in Preschoolers with Autism Spectrum Disorders and Similar Characteristics

    ERIC Educational Resources Information Center

    Ganz, Jennifer B.; Simpson, Richard L.; Corbin-Newsome, Jawanda

    2008-01-01

    By definition children with autism spectrum disorders (ASD) experience difficulty understanding and using language. Accordingly, visual and picture-based strategies such as the Picture Exchange Communication System (PECS) show promise in ameliorating speech and language deficits. This study reports the results of a multiple baseline across…

  1. The Collection and Analysis of Data on Children with Speech, Language and Communication Needs: The Challenge to Education and Health Services

    ERIC Educational Resources Information Center

    Lindsay, Geoff

    2011-01-01

    The Bercow Report (Bercow, 2008) commissioned by the UK government provided a high status impetus to improve services for children and young people with the full range of speech, language and communication needs (SLCN). A research study commissioned to provide evidence to Bercow (2008) identified both limitations and potential benefits regarding…

  2. A Further Comparison of Manual Signing, Picture Exchange, and Speech-Generating Devices as Communication Modes for Children with Autism Spectrum Disorders

    ERIC Educational Resources Information Center

    van der Meer, Larah; Sutherland, Dean; O'Reilly, Mark F.; Lancioni, Giulio E.; Sigafoos, Jeff

    2012-01-01

    We compared acquisition of, and preference for, manual signing (MS), picture exchange (PE), and speech-generating devices (SGDs) in four children with autism spectrum disorders (ASD). Intervention was introduced across participants in a non-concurrent multiple-baseline design and acquisition of the three communication modes was compared in an…

  3. "Do We Make Ourselves Clear?" Developing a Social, Emotional and Behavioural Difficulties (SEBD) Support Service's Effectiveness in Detecting and Supporting Children Experiencing Speech, Language and Communication Difficulties (SLCD)

    ERIC Educational Resources Information Center

    Stiles, Matthew

    2013-01-01

    Research has identified a significant relationship between social, emotional and behavioural difficulties (SEBD) and speech, language and communication difficulties (SLCD). However, little has been published regarding the levels of knowledge and skill that practitioners working with pupils experiencing SEBD have in this important area, nor how…

  4. Common cues to emotion in the dynamic facial expressions of speech and song.

    PubMed

    Livingstone, Steven R; Thompson, William F; Wanderley, Marcelo M; Palmer, Caroline

    2015-01-01

    Speech and song are universal forms of vocalization that may share aspects of emotional expression. Research has focused on parallels in acoustic features, overlooking facial cues to emotion. In three experiments, we compared moving facial expressions in speech and song. In Experiment 1, vocalists spoke and sang statements each with five emotions. Vocalists exhibited emotion-dependent movements of the eyebrows and lip corners that transcended speech-song differences. Vocalists' jaw movements were coupled to their acoustic intensity, exhibiting differences across emotion and speech-song. Vocalists' emotional movements extended beyond vocal sound to include large sustained expressions, suggesting a communicative function. In Experiment 2, viewers judged silent videos of vocalists' facial expressions prior to, during, and following vocalization. Emotional intentions were identified accurately for movements during and after vocalization, suggesting that these movements support the acoustic message. Experiment 3 compared emotional identification in voice-only, face-only, and face-and-voice recordings. Emotion judgements for voice-only singing were poorly identified, yet were accurate for all other conditions, confirming that facial expressions conveyed emotion more accurately than the voice in song, yet were equivalent in speech. Collectively, these findings highlight broad commonalities in the facial cues to emotion in speech and song, yet highlight differences in perception and acoustic-motor production.

  5. Toward a semantics for an agent communications language based on speech-acts

    SciTech Connect

    Smith, I.A.; Cohen, P.R.

    1996-12-31

    Systems based on distributed agent architectures require an agent communications language having a clearly defined semantics. This paper demonstrates that a semantics for an agent communications language can be founded on the premise that agents are building, maintaining, and disbanding teams through their performance of communicative acts. This view requires that definitions of basic communicative acts, such as requesting, be recast in terms of the formation of a joint intention - a mental state that has been suggested underlies team behavior. To illustrate these points, a semantics is developed for a number of communication actions that can form and dissolve teams. It is then demonstrated how much of the structure of popular finite-state dialogue models, such as Winograd and Flores` basic conversation for action, follows as a consequence of the logical relationships that are created by the redefined communicative actions.

  6. Two-microphone spatial filtering provides speech reception benefits for cochlear implant users in difficult acoustic environments

    PubMed Central

    Goldsworthy, Raymond L.; Delhorne, Lorraine A.; Desloge, Joseph G.; Braida, Louis D.

    2014-01-01

    This article introduces and provides an assessment of a spatial-filtering algorithm based on two closely-spaced (∼1 cm) microphones in a behind-the-ear shell. The evaluated spatial-filtering algorithm used fast (∼10 ms) temporal-spectral analysis to determine the location of incoming sounds and to enhance sounds arriving from straight ahead of the listener. Speech reception thresholds (SRTs) were measured for eight cochlear implant (CI) users using consonant and vowel materials under three processing conditions: An omni-directional response, a dipole-directional response, and the spatial-filtering algorithm. The background noise condition used three simultaneous time-reversed speech signals as interferers located at 90°, 180°, and 270°. Results indicated that the spatial-filtering algorithm can provide speech reception benefits of 5.8 to 10.7 dB SRT compared to an omni-directional response in a reverberant room with multiple noise sources. Given the observed SRT benefits, coupled with an efficient design, the proposed algorithm is promising as a CI noise-reduction solution. PMID:25096120

  7. [Speech behavior and neurotic personality structure. Is there a neuroses-specific communication behavior in the psychoanalytic situation].

    PubMed

    Goeppert, S

    1975-01-01

    Speech behaviour of the patient and the analyst is considered with regard to the psychoanalytic situation and its regulating factors such as transference and countertransference. It appears that the scientific viewpoint of the researching analyst determines the conception as well as the interpretation of a possible connection between verbal behaviour and neurotic personality structure. If the psychoanalytic situation is regarded as a means to investigate the patient and his symptoms, his verbal behaviour seems to depend on the unconscious motives forming his neurotic structure of personality. On the contrary, if the psychoanalytic situation is regarded as a communicative situation, the verbal behaviour of the patient as well as of the analyst depends rather on their 'here-and-now' transference relationship in a certain phase of the psychoanalytic process.

  8. Effects of Semantic Context and Feedback on Perceptual Learning of Speech Processed through an Acoustic Simulation of a Cochlear Implant

    ERIC Educational Resources Information Center

    Loebach, Jeremy L.; Pisoni, David B.; Svirsky, Mario A.

    2010-01-01

    The effect of feedback and materials on perceptual learning was examined in listeners with normal hearing who were exposed to cochlear implant simulations. Generalization was most robust when feedback paired the spectrally degraded sentences with their written transcriptions, promoting mapping between the degraded signal and its acoustic-phonetic…

  9. A Scoping Review of Interventions to Supplement Spoken Communication for Children with Limited Speech or Language Skills

    PubMed Central

    Costantino, Maria Antonella; Bonati, Maurizio

    2014-01-01

    Background Augmentative and Alternative Communication (AAC) is used for treating children with severe disorders of speech-language production and/or comprehension. Various strategies are used, but research and debate on their efficacy have remained limited to a specific area and have rarely reached the general medical community. Objective To systematically evaluate outcomes of AAC interventions in children with limited speech or language skills. Methods Searches were conducted (up to December 2012) in the MEDLINE, EMBASE, PsycINFO, CINAHL, DARE, and Cochrane Library databases. Furthermore, relevant journals were searched by hand. References from identified studies were examined. Only RCTs were considered. Trial quality was assessed according to a standardized and validated set of criteria. Results Fourteen of 1661 retrieved papers met inclusion criteria. A total of 666 children were included in the review and 7 papers involved only children <5 years old. Papers were of average quality and all but one had been published during the previous 10 years by one of 8 research groups, 5 of which from the United States. Seven studies directly addressed AAC use by children with different disabilities. Seven studies enrolled typically developing children: 5 evaluated the use of AAC technologies by children without disabilities in order to obtain results that could be used to improve interventions in peers with disabilities, and 2 evaluated peers’ attitudes towards children who used AAC. Both interventions and outcome measures varied widely between studies. Overall findings demonstrate the effectiveness of the AAC interventions considered, but the focus on RCTs alone appears too restrictive. Conclusions Solid evidence of the positive effects of AAC interventions in children with severe communication disorders must be generated, and different methods are needed besides RCTs. Moreover, it is important that knowledge, research, and debate extend to the medical community in order

  10. Supporting culturally and linguistically diverse children with speech, language and communication needs: Overarching principles, individual approaches.

    PubMed

    Verdon, Sarah; McLeod, Sharynne; Wong, Sandie

    2015-01-01

    Speech-language pathologists (SLPs) are working with an increasing number of families from culturally and linguistically diverse backgrounds as the world's population continues to become more internationally mobile. The heterogeneity of these diverse populations makes it impossible to identify and document a one size fits all strategy for working with culturally and linguistically diverse families. This paper explores approaches to practice by SLPs identified as specialising in multilingual and multicultural practice in culturally and linguistically diverse contexts from around the world. Data were obtained from ethnographic observation of 14 sites in 5 countries on 4 continents. The sites included hospital settings, university clinics, school-based settings, private practices and Indigenous community-based services. There were 652 individual artefacts collected from the sites which included interview transcripts, photographs, videos, narrative reflections, informal and formal field notes. The data were analysed using Cultural-Historical Activity Theory (Engeström, 1987). From the analysis six overarching Principles of Culturally Competent Practice (PCCP) were identified. These were: (1) identification of culturally appropriate and mutually motivating therapy goals, (2) knowledge of languages and culture, (3) use of culturally appropriate resources, (4) consideration of the cultural, social and political context, (5) consultation with families and communities, and (6) collaboration between professionals. These overarching principles align with the six position statements developed by the International Expert Panel on Multilingual Children's Speech (2012) which aim to enhance the cultural competence of speech pathologists and their practice. The international examples provided in the current study demonstrate the individualised ways that these overarching principles are enacted in a range of different organisational, social, cultural and political contexts

  11. Supporting culturally and linguistically diverse children with speech, language and communication needs: Overarching principles, individual approaches.

    PubMed

    Verdon, Sarah; McLeod, Sharynne; Wong, Sandie

    2015-01-01

    Speech-language pathologists (SLPs) are working with an increasing number of families from culturally and linguistically diverse backgrounds as the world's population continues to become more internationally mobile. The heterogeneity of these diverse populations makes it impossible to identify and document a one size fits all strategy for working with culturally and linguistically diverse families. This paper explores approaches to practice by SLPs identified as specialising in multilingual and multicultural practice in culturally and linguistically diverse contexts from around the world. Data were obtained from ethnographic observation of 14 sites in 5 countries on 4 continents. The sites included hospital settings, university clinics, school-based settings, private practices and Indigenous community-based services. There were 652 individual artefacts collected from the sites which included interview transcripts, photographs, videos, narrative reflections, informal and formal field notes. The data were analysed using Cultural-Historical Activity Theory (Engeström, 1987). From the analysis six overarching Principles of Culturally Competent Practice (PCCP) were identified. These were: (1) identification of culturally appropriate and mutually motivating therapy goals, (2) knowledge of languages and culture, (3) use of culturally appropriate resources, (4) consideration of the cultural, social and political context, (5) consultation with families and communities, and (6) collaboration between professionals. These overarching principles align with the six position statements developed by the International Expert Panel on Multilingual Children's Speech (2012) which aim to enhance the cultural competence of speech pathologists and their practice. The international examples provided in the current study demonstrate the individualised ways that these overarching principles are enacted in a range of different organisational, social, cultural and political contexts

  12. Tutorial on architectural acoustics

    NASA Astrophysics Data System (ADS)

    Shaw, Neil; Talaske, Rick; Bistafa, Sylvio

    2002-11-01

    This tutorial is intended to provide an overview of current knowledge and practice in architectural acoustics. Topics covered will include basic concepts and history, acoustics of small rooms (small rooms for speech such as classrooms and meeting rooms, music studios, small critical listening spaces such as home theatres) and the acoustics of large rooms (larger assembly halls, auditoria, and performance halls).

  13. Wake-up-word speech recognition application for first responder communication enhancement

    NASA Astrophysics Data System (ADS)

    Këpuska, Veton; Breitfeller, Jason

    2006-05-01

    Speech Recognition systems, historically, have proven to be cumbersome and insufficiently accurate for a range of applications. The ultimate goal of our proposed technology is to fundamentally change the way current Speech Recognition (SR) systems interact with humans and develop an application that is extremely hardware efficient. Accurate SR and reasonable hardware requirements will afford the average first responder officer, e.g., police officer, a true break-through technology that will change the way an officer performs his duties. The presented technology provides a cutting-edge solution for human-machine interaction through the utilization of a properly solved Wake-Up-Word (WUW) SR problem. This paradigm-shift provides the basis for development of SR systems with truly "Voice Activated" capabilities, impacting all SR based technologies and the way in which humans interact with computers. This shift is a radical departure from the current "push-to-talk" paradigm currently applied to all speech-to-text or speech-recognition applications. To be able to achieve this goal, a significantly more accurate pattern classification and scoring technique is required, which in turn provides SR systems enhanced performance for correct recognition (i.e., minimization of false rejection) as well as correct rejection (i.e., minimization of false acceptance). A revolutionary and innovative classification and scoring technique is used that is a significant enhancement over an earlier method presented in reference [1]. The solution in reference [1] has been demonstrated to meet the stringent requirements of the WUW-SR task. Advanced solution of [1] is a novel technique that is model and algorithm independent. Therefore, it could be used to significantly improve performance of existing recognition algorithms and systems. Reduction of error rates of over 40% are commonly observed for both false rejections and false acceptance. In this paper the architecture of the WUW-SR based

  14. Sign Language, Speech, and Communication Repair Abilities by Children with Congenital Deafblindness

    ERIC Educational Resources Information Center

    Cascella, Paul W.; Bruce, Susan M.; Trief, Ellen

    2015-01-01

    There are many reports that identify the communication of persons with deafblindness. These reports are descriptive in nature, highlighting comprehension, gesture development, communication actions (that is, forms) and purposes (that is, functions), and the collection of verbal skills. This report extends the literature with an in-depth…

  15. Acoustic communication in two freshwater gobies: the relationship between ambient noise, hearing thresholds and sound spectrum.

    PubMed

    Lugli, M; Yan, H Y; Fine, M L

    2003-04-01

    Two freshwater gobies Padogobius martensii and Gobius nigricans live in shallow (5-70 cm) stony streams, and males of both species produce courtship sounds. A previous study demonstrated high noise levels near waterfalls, a quiet window in the noise around 100 Hz at noisy locations, and extremely short-range propagation of noise and goby signals. To investigate the relationship of this acoustic environment to communication, we determined audiograms for both species and measured parameters of courtship sounds produced in the streams. We also deflated the swimbladder in P. martensii to determine its effect on frequency utilization in sound production and hearing. Both species are maximally sensitive at 100 Hz and produce low-frequency sounds with main energy from 70 to 100-150 Hz. Swimbladder deflation does not affect auditory threshold or dominant frequency of courtship sounds and has no or minor effects on sound amplitude. Therefore, both species utilize frequencies for hearing and sound production that fall within the low-frequency quiet region, and the equivalent relationship between auditory sensitivity and maximum ambient noise levels in both species further suggests that ambient noise shapes hearing sensitivity. PMID:12665991

  16. Acoustic communication in two freshwater gobies: the relationship between ambient noise, hearing thresholds and sound spectrum.

    PubMed

    Lugli, M; Yan, H Y; Fine, M L

    2003-04-01

    Two freshwater gobies Padogobius martensii and Gobius nigricans live in shallow (5-70 cm) stony streams, and males of both species produce courtship sounds. A previous study demonstrated high noise levels near waterfalls, a quiet window in the noise around 100 Hz at noisy locations, and extremely short-range propagation of noise and goby signals. To investigate the relationship of this acoustic environment to communication, we determined audiograms for both species and measured parameters of courtship sounds produced in the streams. We also deflated the swimbladder in P. martensii to determine its effect on frequency utilization in sound production and hearing. Both species are maximally sensitive at 100 Hz and produce low-frequency sounds with main energy from 70 to 100-150 Hz. Swimbladder deflation does not affect auditory threshold or dominant frequency of courtship sounds and has no or minor effects on sound amplitude. Therefore, both species utilize frequencies for hearing and sound production that fall within the low-frequency quiet region, and the equivalent relationship between auditory sensitivity and maximum ambient noise levels in both species further suggests that ambient noise shapes hearing sensitivity.

  17. Relating the performance of time-reversal-based underwater acoustic communications in different shallow water environments.

    PubMed

    Yang, T C

    2011-10-01

    The performance of underwater acoustic communications, such as the output signal-to-noise ratio (OSNR), is generally dependent on the channel specifics, hence a channel model is normally required as the performance of the channel equalizer depends on the number of tap coefficients used (e.g., a sparse equalizer) which are different for different oceans having different multipath arrivals. This letter presents theoretical arguments, and experimental data from different oceans that suggest that the increase of OSNR with the number of diverse receivers (in terms of the effective number of receivers) and the decrease of OSNR with the channel-estimation error follow a universal relationship using the time-reversal or correlation-based equalizer, despite the fact that the channels have very different properties. The reason is due to the fact that the OSNR is a function of the q function, the auto-correlation of the received impulse responses summed over all receiver channels, and the q function is approximately the same for all shallow waters given a sufficient (≥4-6) number of receivers.

  18. The African cichlid fish Astatotilapia burtoni uses acoustic communication for reproduction: sound production, hearing, and behavioral significance.

    PubMed

    Maruska, Karen P; Ung, Uyhun S; Fernald, Russell D

    2012-01-01

    Sexual reproduction in all animals depends on effective communication between signalers and receivers. Many fish species, especially the African cichlids, are well known for their bright coloration and the importance of visual signaling during courtship and mate choice, but little is known about what role acoustic communication plays during mating and how it contributes to sexual selection in this phenotypically diverse group of vertebrates. Here we examined acoustic communication during reproduction in the social cichlid fish, Astatotilapia burtoni. We characterized the sounds and associated behaviors produced by dominant males during courtship, tested for differences in hearing ability associated with female reproductive state and male social status, and then tested the hypothesis that female mate preference is influenced by male sound production. We show that dominant males produce intentional courtship sounds in close proximity to females, and that sounds are spectrally similar to their hearing abilities. Females were 2-5-fold more sensitive to low frequency sounds in the spectral range of male courtship sounds when they were sexually-receptive compared to during the mouthbrooding parental phase. Hearing thresholds were also negatively correlated with circulating sex-steroid levels in females but positively correlated in males, suggesting a potential role for steroids in reproductive-state auditory plasticity. Behavioral experiments showed that receptive females preferred to affiliate with males that were associated with playback of courtship sounds compared to noise controls, indicating that acoustic information is likely important for female mate choice. These data show for the first time in a Tanganyikan cichlid that acoustic communication is important during reproduction as part of a multimodal signaling repertoire, and that perception of auditory information changes depending on the animal's internal physiological state. Our results highlight the

  19. A user's guide for the signal processing software for image and speech compression developed in the Communications and Signal Processing Laboratory (CSPL), version 1

    NASA Technical Reports Server (NTRS)

    Kumar, P.; Lin, F. Y.; Vaishampayan, V.; Farvardin, N.

    1986-01-01

    A complete documentation of the software developed in the Communication and Signal Processing Laboratory (CSPL) during the period of July 1985 to March 1986 is provided. Utility programs and subroutines that were developed for a user-friendly image and speech processing environment are described. Additional programs for data compression of image and speech type signals are included. Also, programs for the zero-memory and block transform quantization in the presence of channel noise are described. Finally, several routines for simulating the perfromance of image compression algorithms are included.

  20. Machine Translation from Speech

    NASA Astrophysics Data System (ADS)

    Schwartz, Richard; Olive, Joseph; McCary, John; Christianson, Caitlin

    This chapter describes approaches for translation from speech. Translation from speech presents two new issues. First, of course, we must recognize the speech in the source language. Although speech recognition has improved considerably over the last three decades, it is still far from being a solved problem. In the best of conditions, when the speech comes from high quality, carefully enunciated speech, on common topics (such as speech read by a trained news broadcaster), the word error rate is typically on the order of 5%. Humans can typically transcribe speech like this with less than 1% disagreement between annotators, so even this best number is still far worse than human performance. However, the task gets much harder when anything changes from this ideal condition. Some of the conditions that cause higher error rate are, if the topic is somewhat unusual, or the speakers are not reading so that their speech is more spontaneous, or if the speakers have an accent or are speaking a dialect, or if there is any acoustic degradation, such as noise or reverberation. In these cases, the word error can increase significantly to 20%, 30%, or higher. Accordingly, most of this chapter discusses techniques for improving speech recognition accuracy, while one section discusses techniques for integrating speech recognition with translation.