Psychoacoustic entropy theory and its implications for performance practice
NASA Astrophysics Data System (ADS)
Strohman, Gregory J.
This dissertation attempts to motivate, derive and imply potential uses for a generalized perceptual theory of musical harmony called psychoacoustic entropy theory. This theory treats the human auditory system as a physical system which takes acoustic measurements. As a result, the human auditory system is subject to all the appropriate uncertainties and limitations of other physical measurement systems. This is the theoretic basis for defining psychoacoustic entropy. Psychoacoustic entropy is a numerical quantity which indexes the degree to which the human auditory system perceives instantaneous disorder within a sound pressure wave. Chapter one explains the importance of harmonic analysis as a tool for performance practice. It also outlines the critical limitations for many of the most influential historical approaches to modeling harmonic stability, particularly when compared to available scientific research in psychoacoustics. Rather than analyze a musical excerpt, psychoacoustic entropy is calculated directly from sound pressure waves themselves. This frames psychoacoustic entropy theory in the most general possible terms as a theory of musical harmony, enabling it to be invoked for any perceivable sound. Chapter two provides and examines many widely accepted mathematical models of the acoustics and psychoacoustics of these sound pressure waves. Chapter three introduces entropy as a precise way of measuring perceived uncertainty in sound pressure waves. Entropy is used, in combination with the acoustic and psychoacoustic models introduced in chapter two, to motivate the mathematical formulation of psychoacoustic entropy theory. Chapter four shows how to use psychoacoustic entropy theory to analyze the certain types of musical harmonies, while chapter five applies the analytical tools developed in chapter four to two short musical excerpts to influence their interpretation. Almost every form of harmonic analysis invokes some degree of mathematical reasoning. However, the limited scope of most harmonic systems used for Western common practice music greatly simplifies the necessary level of mathematical detail. Psychoacoustic entropy theory requires a greater deal of mathematical complexity due to its sheer scope as a generalized theory of musical harmony. Fortunately, under specific assumptions the theory can take on vastly simpler forms. Psychoacoustic entropy theory appears to be highly compatible with the latest scientific research in psychoacoustics. However, the theory itself should be regarded as a hypothesis and this dissertation an experiment in progress. The evaluation of psychoacoustic entropy theory as a scientific theory of human sonic perception must wait for more rigorous future research.
Peñaloza López, Yolanda Rebeca; Orozco Peña, Xóchitl Daisy; Pérez Ruiz, Santiago Jesús
2018-04-03
To evaluate the central auditory processing disorders in patients with multiple sclerosis, emphasizing auditory laterality by applying psychoacoustic tests and to identify their relationship with the Multiple Sclerosis Disability Scale (EDSS) functions. Depression scales (HADS), EDSS, and 9 psychoacoustic tests to study CAPD were applied to 26 individuals with multiple sclerosis and 26 controls. Correlation tests were performed between the EDSS and psychoacoustic tests. Seven out of 9 psychoacoustic tests were significantly different (P<.05); right or left (14/19 explorations) with respect to control. In dichotic digits there was a left-ear advantage compared to the usual predominance of RDD. There was significant correlation in five psychoacoustic tests and the specific functions of EDSS. The left-ear advantage detected and interpreted as an expression of deficient influences of the corpus callosum and attention in multiple sclerosis should be investigated. There was a correlation between psychoacoustic tests and specific EDSS functions. Copyright © 2018 Sociedad Española de Otorrinolaringología y Cirugía de Cabeza y Cuello. Publicado por Elsevier España, S.L.U. All rights reserved.
Choice and Effects of Instrument Sound in Aural Training
ERIC Educational Resources Information Center
Loh, Christian Sebastian
2007-01-01
A musical note produced through the vibration of a single string is psychoacoustically simpler/purer than that produced via multiple-strings vibration. Does the psychoacoustics of instrument sound have any effect on learning outcomes in music instruction? This study investigated the effect of two psychoacoustically distinct instrument sounds on…
Uncertainty Estimates of Psychoacoustic Thresholds Obtained from Group Tests
NASA Technical Reports Server (NTRS)
Rathsam, Jonathan; Christian, Andrew
2016-01-01
Adaptive psychoacoustic test methods, in which the next signal level depends on the response to the previous signal, are the most efficient for determining psychoacoustic thresholds of individual subjects. In many tests conducted in the NASA psychoacoustic labs, the goal is to determine thresholds representative of the general population. To do this economically, non-adaptive testing methods are used in which three or four subjects are tested at the same time with predetermined signal levels. This approach requires us to identify techniques for assessing the uncertainty in resulting group-average psychoacoustic thresholds. In this presentation we examine the Delta Method of frequentist statistics, the Generalized Linear Model (GLM), the Nonparametric Bootstrap, a frequentist method, and Markov Chain Monte Carlo Posterior Estimation and a Bayesian approach. Each technique is exercised on a manufactured, theoretical dataset and then on datasets from two psychoacoustics facilities at NASA. The Delta Method is the simplest to implement and accurate for the cases studied. The GLM is found to be the least robust, and the Bootstrap takes the longest to calculate. The Bayesian Posterior Estimate is the most versatile technique examined because it allows the inclusion of prior information.
NASA Astrophysics Data System (ADS)
Sudarsono, Anugrah S.; Merthayasa, I. G. N.; Suprijanto
2015-09-01
This research tried to compare psycho-acoustics and Physio-acoustic measurement to find the optimum reverberation time of soundfield from angklung music. Psycho-acoustic measurement was conducted using a paired comparison method and Physio-acoustic measurement was conducted with EEG Measurement on T3, T4, FP1, and FP2 measurement points. EEG measurement was conducted with 5 persons. Pentatonic angklung music was used as a stimulus with reverberation time variation. The variation was between 0.8 s - 1.6 s with 0.2 s step. EEG signal was analysed using a Power Spectral Density method on Alpha Wave, High Alpha Wave, and Theta Wave. Psycho-acoustic measurement on 50 persons showed that reverberation time preference of pentatonic angklung music was 1.2 second. The result was similar to Theta Wave measurement on FP2 measurement point. High Alpha wave on T4 measurement gave different results, but had similar patterns with psycho-acoustic measurement
Psychoacoustical evaluation of natural and urban sounds in soundscapes.
Yang, Ming; Kang, Jian
2013-07-01
Among various sounds in the environment, natural sounds, such as water sounds and birdsongs, have proven to be highly preferred by humans, but the reasons for these preferences have not been thoroughly researched. This paper explores differences between various natural and urban environmental sounds from the viewpoint of objective measures, especially psychoacoustical parameters. The sound samples used in this study include the recordings of single sound source categories of water, wind, birdsongs, and urban sounds including street music, mechanical sounds, and traffic noise. The samples are analyzed with a number of existing psychoacoustical parameter algorithmic models. Based on hierarchical cluster and principal components analyses of the calculated results, a series of differences has been shown among different sound types in terms of key psychoacoustical parameters. While different sound categories cannot be identified using any single acoustical and psychoacoustical parameter, identification can be made with a group of parameters, as analyzed with artificial neural networks and discriminant functions in this paper. For artificial neural networks, correlations between network predictions and targets using the average and standard deviation data of psychoacoustical parameters as inputs are above 0.95 for the three natural sound categories and above 0.90 for the urban sound category. For sound identification/classification, key parameters are fluctuation strength, loudness, and sharpness.
Chowdhury, Nafees Uddin; Otomaru, Takafumi; Murase, Mai; Inohara, Ken; Hattori, Mariko; Sumita, Yuka I; Taniguchi, Hisashi
2011-01-01
An objective assessment of speech would benefit the prosthetic rehabilitation of maxillectomy patients. This study aimed to establish a simple, objective evaluation of monosyllable /sa/ utterances in maxillectomy patients by using a psychoacoustic system typically used in industry. This study comprised two experiments. Experiment 1 involved analysis of the psychoacoustic parameters (loudness, sharpness and roughness) in monosyllable /sa/ utterances by 18 healthy subjects (9 males, 9 females). The utterances were recorded in a sound-treated room. The coefficient of variation (CV) for each parameter was compared to identify the most suitable parameter for objective evaluation of speech. Experiment 2 involved analysis of /sa/ utterances by 18 maxillectomy patients (9 males, 9 females) with and without prosthesis, and comparisons of the psychoacoustic data between the healthy subjects and maxillectomy patients without prosthesis, between the maxillectomy patients with and without prosthesis, and between the healthy subjects and maxillectomy patients with prosthesis. The CV for sharpness was the lowest among the three psychoacoustic parameters in both the healthy males and females. There were significant differences in the sharpness of /sa/ between the healthy subjects and the maxillectomy patients without prosthesis (but not with prosthesis), and between the maxillectomy patients with and without prosthesis. We found that the psychoacoustic parameters typically adopted in industrial research could also be applied to evaluate the psychoacoustics of the monosyllable /sa/ utterance, and distinguished the monosyllable /sa/ in maxillectomy patients with an obturator from that without an obturator using the system. Copyright © 2010 Japan Prosthodontic Society. Published by Elsevier Ltd. All rights reserved.
Influence of musical and psychoacoustical training on pitch discrimination.
Micheyl, Christophe; Delhommeau, Karine; Perrot, Xavier; Oxenham, Andrew J
2006-09-01
This study compared the influence of musical and psychoacoustical training on auditory pitch discrimination abilities. In a first experiment, pitch discrimination thresholds for pure and complex tones were measured in 30 classical musicians and 30 non-musicians, none of whom had prior psychoacoustical training. The non-musicians' mean thresholds were more than six times larger than those of the classical musicians initially, and still about four times larger after 2h of training using an adaptive two-interval forced-choice procedure; this difference is two to three times larger than suggested by previous studies. The musicians' thresholds were close to those measured in earlier psychoacoustical studies using highly trained listeners, and showed little improvement with training; this suggests that classical musical training can lead to optimal or nearly optimal pitch discrimination performance. A second experiment was performed to determine how much additional training was required for the non-musicians to obtain thresholds as low as those of the classical musicians from experiment 1. Eight new non-musicians with no prior training practiced the frequency discrimination task for a total of 14 h. It took between 4 and 8h of training for their thresholds to become as small as those measured in the classical musicians from experiment 1. These findings supplement and qualify earlier data in the literature regarding the respective influence of musical and psychoacoustical training on pitch discrimination performance.
Multiple Looks in Speech Sound Discrimination in Adults
ERIC Educational Resources Information Center
Holt, Rachael Frush; Carney, Arlene Earley
2005-01-01
N. F. Viemeister and G. H. Wakefield's (1991) multiple looks hypothesis is a theoretical approach from the psychoacoustic literature that has promise for bridging the gap between results from speech perception research and results from psychoacoustic research. This hypothesis accounts for sensory detection data and predicts that if the "looks" at…
PSYCHOACOUSTICS: a comprehensive MATLAB toolbox for auditory testing.
Soranzo, Alessandro; Grassi, Massimo
2014-01-01
PSYCHOACOUSTICS is a new MATLAB toolbox which implements three classic adaptive procedures for auditory threshold estimation. The first includes those of the Staircase family (method of limits, simple up-down and transformed up-down); the second is the Parameter Estimation by Sequential Testing (PEST); and the third is the Maximum Likelihood Procedure (MLP). The toolbox comes with more than twenty built-in experiments each provided with the recommended (default) parameters. However, if desired, these parameters can be modified through an intuitive and user friendly graphical interface and stored for future use (no programming skills are required). Finally, PSYCHOACOUSTICS is very flexible as it comes with several signal generators and can be easily extended for any experiment.
PSYCHOACOUSTICS: a comprehensive MATLAB toolbox for auditory testing
Soranzo, Alessandro; Grassi, Massimo
2014-01-01
PSYCHOACOUSTICS is a new MATLAB toolbox which implements three classic adaptive procedures for auditory threshold estimation. The first includes those of the Staircase family (method of limits, simple up-down and transformed up-down); the second is the Parameter Estimation by Sequential Testing (PEST); and the third is the Maximum Likelihood Procedure (MLP). The toolbox comes with more than twenty built-in experiments each provided with the recommended (default) parameters. However, if desired, these parameters can be modified through an intuitive and user friendly graphical interface and stored for future use (no programming skills are required). Finally, PSYCHOACOUSTICS is very flexible as it comes with several signal generators and can be easily extended for any experiment. PMID:25101013
Sato, Shin-ichi; You, Jin; Jeon, Jin Yong
2007-07-01
Psychoacoustical and autocorrelation function (ACF) parameters were employed to describe the temporal fluctuations of refrigerator noise during starting, transition into/from the stationary phase and termination of operation. The temporal fluctuations of refrigerator noise include a click at start-up, followed by a rapid increase in volume, a change of pitch, and termination of the operation. Subjective evaluations of the noise of 24 different refrigerators were conducted in a real living environment. The relationship between objective measures and perceived noisiness was examined by multiple regression analysis. Sound quality indices were developed based on psychoacoustical and ACF parameters. The psychoacoustical parameters found to be important for evaluating noisiness in the stationary phase were loudness and roughness. The relationship between noisiness and ACF parameters shows that sound energy and its fluctuations are important for evaluating noisiness. Also, refrigerator sounds that had a fluctuation of pitch were rated as more annoying. The tolerance level for the starting phase of refrigerator noise was found to be 33 dBA, which is the level where 65% of the participants in the subjective tests were satisfied.
Tsakiraki, Eleni S; Tsiaparas, Nikolaos N; Christopoulou, Maria I; Papageorgiou, Charalabos Ch; Nikita, Konstantina S
2014-01-01
The aim of the paper is the assessment of neural potentials disorder during a differential sensitivity psychoacoustic procedure. Ten volunteers were asked to compare the duration of two acoustic pulses: one reference with stable duration of 500 ms and one trial which varied from 420 ms to 620 ms. During the discrimination task, Electroencephalogram (EEG) and Event Related Potential (ERP) signals were recorded. The mean Relative Wavelet Energy (mRWE) and the normalized Shannon Wavelet Entropy (nSWE) are computed based on the Discrete Wavelet analysis. The results are correlated to the data derived by the psychoacoustic analysis on the volunteers responses. In most of the electrodes, when the duration of the trial pulse is 460 ms and 560 ms, there is an increase and a decrease in nSWE value, respectively, which is determined mostly by the mRWE in delta rhythm. These extrema are correlated to the Just Noticeable Difference (JND) in pulses duration, calculated by psychoacoustic analysis. The dominance of delta rhythm during the whole auditory experiment is noteworthy. The lowest values of nSWE are noted in temporal lobe.
Kricos, Patricia B.
2006-01-01
The number and proportion of older adults in the United States population is increasing, and more clinical audiologists will be called upon to deliver hearing care to the approximately 35% to 50% of them who experience hearing difficulties. In recent years, the characteristics and sources of receptive communication difficulties in older individuals have been investigated by hearing scientists, cognitive psychologists, and audiologists. It is becoming increasingly apparent that cognitive compromises and psychoacoustic auditory processing disorders associated with aging may contribute to communication difficulties in this population. This paper presents an overview of best practices, based on our current knowledge base, for clinical management of older individuals with limitations in cognitive or psychoacoustic auditory processing capabilities, or both, that accompany aging. PMID:16528428
Díaz-Leines, Sergio; Peñaloza-López, Yolanda R; Serrano-Miranda, Tirzo A; Flores-Ávalos, Blanca; Vidal-Ixta, Martha T; Jiménez-Herrera, Blanca
2013-01-01
Hyperhomocysteinemia as a risk factor for hearing impairment, neuronal damage and cognitive impairment in elderly patients is controversial and is limited by the small number of studies. The aim of this work was determine if elderly patients detected with hyperhomocysteinemia have an increased risk of developing abnormalities in the central auditory processes as compared with a group of patients with appropriate homocysteine levels, and to define the behaviour of psychoacoustic tests and long latency potentials (P300) in these patients. This was a cross-sectional, comparative and analytical study. We formed a group of patients with hyperhomocysteinemia and a control group with normal levels of homocysteine. All patients underwent audiometry, tympanometry and a selection of psychoacoustic tests (dichotic digits, low-pass filtered words, speech in noise and masking level difference), auditory evoked brainstem potentials and P300. Patients with hyperhomocysteinemia had higher values in the test of masking level difference than did the control group (P=.049) and more protracted latency in P300 (P=.000). Hyperhomocysteinemia is a factor that alters the central auditory functions. Alterations in psychoacoustic tests and disturbances in electrophysiological tests suggest that the central portion of the auditory pathway is affected in patients with hyperhomocysteinemia. Copyright © 2012 Elsevier España, S.L. All rights reserved.
NASA Astrophysics Data System (ADS)
Oxenham, Andrew J.; Rosengard, Peninah S.; Braida, Louis D.
2004-05-01
Cochlear damage can lead to a reduction in the overall amount of peripheral auditory compression, presumably due to outer hair cell (OHC) loss or dysfunction. The perceptual consequences of functional OHC loss include loudness recruitment and reduced dynamic range, poorer frequency selectivity, and poorer effective temporal resolution. These in turn may lead to a reduced ability to make use of spectral and temporal fluctuations in background noise when listening to a target sound, such as speech. We tested the effect of OHC function on speech reception in hearing-impaired listeners by comparing psychoacoustic measures of cochlear compression and sentence recognition in a variety of noise backgrounds. In line with earlier studies, we found weak (nonsignificant) correlations between the psychoacoustic tasks and speech reception thresholds in quiet or in steady-state noise. However, when spectral and temporal fluctuations were introduced in the masker, speech reception improved to an extent that was well predicted by the psychoacoustic measures. Thus, our initial results suggest a strong relationship between measures of cochlear compression and the ability of listeners to take advantage of spectral and temporal masker fluctuations in recognizing speech. [Work supported by NIH Grants Nos. R01DC03909, T32DC00038, and R01DC00117.
Crowell, Sara E.; Wells-Berlin, Alicia M.; Therrien, Ronald E.; Yannuzzi, Sally E.; Carr, Catherine E.
2016-01-01
Auditory sensitivity was measured in a species of diving duck that is not often kept in captivity, the lesser scaup. Behavioral (psychoacoustics) and electrophysiological [the auditory brainstem response (ABR)] methods were used to measure in-air auditory sensitivity, and the resulting audiograms were compared. Both approaches yielded audiograms with similar U-shapes and regions of greatest sensitivity (2000−3000 Hz). However, ABR thresholds were higher than psychoacoustic thresholds at all frequencies. This difference was least at the highest frequency tested using both methods (5700 Hz) and greatest at 1000 Hz, where the ABR threshold was 26.8 dB higher than the behavioral measure of threshold. This difference is commonly reported in studies involving many different species. These results highlight the usefulness of each method, depending on the testing conditions and availability of the animals.
Crowell, Sara E; Wells-Berlin, Alicia M; Therrien, Ronald E; Yannuzzi, Sally E; Carr, Catherine E
2016-05-01
Auditory sensitivity was measured in a species of diving duck that is not often kept in captivity, the lesser scaup. Behavioral (psychoacoustics) and electrophysiological [the auditory brainstem response (ABR)] methods were used to measure in-air auditory sensitivity, and the resulting audiograms were compared. Both approaches yielded audiograms with similar U-shapes and regions of greatest sensitivity (2000-3000 Hz). However, ABR thresholds were higher than psychoacoustic thresholds at all frequencies. This difference was least at the highest frequency tested using both methods (5700 Hz) and greatest at 1000 Hz, where the ABR threshold was 26.8 dB higher than the behavioral measure of threshold. This difference is commonly reported in studies involving many different species. These results highlight the usefulness of each method, depending on the testing conditions and availability of the animals.
3-D Sound for Virtual Reality and Multimedia
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Trejo, Leonard J. (Technical Monitor)
2000-01-01
Technology and applications for the rendering of virtual acoustic spaces are reviewed. Chapter 1 deals with acoustics and psychoacoustics. Chapters 2 and 3 cover cues to spatial hearing and review psychoacoustic literature. Chapter 4 covers signal processing and systems overviews of 3-D sound systems. Chapter 5 covers applications to computer workstations, communication systems, aeronautics and space, and sonic arts. Chapter 6 lists resources. This TM is a reprint of the 1994 book from Academic Press.
Mo, Lei
2017-01-01
The scientific community has been divided as to the origin of individual differences in perceiving the sounds of a second language (L2). There are two alternative explanations: a general psychoacoustic origin vs. a speech-specific one. A previous study showed that such individual variability is linked to the perceivers’ speech-specific capabilities, rather than the perceivers’ psychoacoustic abilities. However, we assume that the selection of participants and parameters of sound stimuli might not appropriate. Therefore, we adjusted the sound stimuli and recorded event-related potentials (ERPs) from two groups of early, proficient Cantonese (L1)-Mandarin (L2) bilinguals who differed in their mastery of the Mandarin (L2) phonetic contrast /in-ing/, to explore whether the individual differences in perceiving L2 stem from participants’ ability to discriminate various pure tones (frequency, duration and pattern). To precisely measure the participants’ acoustic discrimination, mismatch negativity (MMN) elicited by the oddball paradigm was recorded in the experiment. The results showed that significant differences between good perceivers (GPs) and poor perceivers (PPs) were found in the three general acoustic conditions (frequency, duration and pattern), and the MMN amplitude for GP was significantly larger than for PP. Therefore, our results support a general psychoacoustic origin of individual variability in L2 phonetic mastery. PMID:29176886
Lin, Yi; Fan, Ruolin; Mo, Lei
2017-01-01
The scientific community has been divided as to the origin of individual differences in perceiving the sounds of a second language (L2). There are two alternative explanations: a general psychoacoustic origin vs. a speech-specific one. A previous study showed that such individual variability is linked to the perceivers' speech-specific capabilities, rather than the perceivers' psychoacoustic abilities. However, we assume that the selection of participants and parameters of sound stimuli might not appropriate. Therefore, we adjusted the sound stimuli and recorded event-related potentials (ERPs) from two groups of early, proficient Cantonese (L1)-Mandarin (L2) bilinguals who differed in their mastery of the Mandarin (L2) phonetic contrast /in-ing/, to explore whether the individual differences in perceiving L2 stem from participants' ability to discriminate various pure tones (frequency, duration and pattern). To precisely measure the participants' acoustic discrimination, mismatch negativity (MMN) elicited by the oddball paradigm was recorded in the experiment. The results showed that significant differences between good perceivers (GPs) and poor perceivers (PPs) were found in the three general acoustic conditions (frequency, duration and pattern), and the MMN amplitude for GP was significantly larger than for PP. Therefore, our results support a general psychoacoustic origin of individual variability in L2 phonetic mastery.
Lopez Valdes, Alejandro; Mc Laughlin, Myles; Viani, Laura; Walshe, Peter; Smith, Jaclyn; Zeng, Fan-Gang; Reilly, Richard B.
2014-01-01
Cochlear implants (CIs) can partially restore functional hearing in deaf individuals. However, multiple factors affect CI listener's speech perception, resulting in large performance differences. Non-speech based tests, such as spectral ripple discrimination, measure acoustic processing capabilities that are highly correlated with speech perception. Currently spectral ripple discrimination is measured using standard psychoacoustic methods, which require attentive listening and active response that can be difficult or even impossible in special patient populations. Here, a completely objective cortical evoked potential based method is developed and validated to assess spectral ripple discrimination in CI listeners. In 19 CI listeners, using an oddball paradigm, cortical evoked potential responses to standard and inverted spectrally rippled stimuli were measured. In the same subjects, psychoacoustic spectral ripple discrimination thresholds were also measured. A neural discrimination threshold was determined by systematically increasing the number of ripples per octave and determining the point at which there was no longer a significant difference between the evoked potential response to the standard and inverted stimuli. A correlation was found between the neural and the psychoacoustic discrimination thresholds (R2 = 0.60, p<0.01). This method can objectively assess CI spectral resolution performance, providing a potential tool for the evaluation and follow-up of CI listeners who have difficulty performing psychoacoustic tests, such as pediatric or new users. PMID:24599314
Lopez Valdes, Alejandro; Mc Laughlin, Myles; Viani, Laura; Walshe, Peter; Smith, Jaclyn; Zeng, Fan-Gang; Reilly, Richard B
2014-01-01
Cochlear implants (CIs) can partially restore functional hearing in deaf individuals. However, multiple factors affect CI listener's speech perception, resulting in large performance differences. Non-speech based tests, such as spectral ripple discrimination, measure acoustic processing capabilities that are highly correlated with speech perception. Currently spectral ripple discrimination is measured using standard psychoacoustic methods, which require attentive listening and active response that can be difficult or even impossible in special patient populations. Here, a completely objective cortical evoked potential based method is developed and validated to assess spectral ripple discrimination in CI listeners. In 19 CI listeners, using an oddball paradigm, cortical evoked potential responses to standard and inverted spectrally rippled stimuli were measured. In the same subjects, psychoacoustic spectral ripple discrimination thresholds were also measured. A neural discrimination threshold was determined by systematically increasing the number of ripples per octave and determining the point at which there was no longer a significant difference between the evoked potential response to the standard and inverted stimuli. A correlation was found between the neural and the psychoacoustic discrimination thresholds (R2=0.60, p<0.01). This method can objectively assess CI spectral resolution performance, providing a potential tool for the evaluation and follow-up of CI listeners who have difficulty performing psychoacoustic tests, such as pediatric or new users.
Balkenhol, Tobias; Wallhäusser-Franke, Elisabeth; Delb, Wolfgang
2013-01-01
Background The phantom auditory perception of subjective tinnitus is associated with aberrant brain activity as evidenced by magneto- and electroencephalographic studies. We tested the hypotheses (1) that psychoacoustically measured tinnitus loudness is related to gamma oscillatory band power, and (2) that tinnitus loudness and tinnitus-related distress are related to distinct brain activity patterns as suggested by the distinction between loudness and distress experienced by tinnitus patients. Furthermore, we explored (3) how hearing impairment, minimum masking level, and (4) psychological comorbidities are related to spontaneous oscillatory brain activity in tinnitus patients. Methods and Findings Resting state oscillatory brain activity recorded electroencephalographically from 46 male tinnitus patients showed a positive correlation between gamma band oscillations and psychoacoustic tinnitus loudness determined with the reconstructed tinnitus sound, but not with the other psychoacoustic loudness measures that were used. Tinnitus-related distress did also correlate with delta band activity, but at electrode positions different from those associated with tinnitus loudness. Furthermore, highly distressed tinnitus patients exhibited a higher level of theta band activity. Moreover, mean hearing loss between 0.125 kHz and 16 kHz was associated with a decrease in gamma activity, whereas minimum masking levels correlated positively with delta band power. In contrast, psychological comorbidities did not express significant correlations with oscillatory brain activity. Conclusion Different clinically relevant tinnitus characteristics show distinctive associations with spontaneous brain oscillatory power. Results support hypothesis (1), but exclusively for the tinnitus loudness derived from matching to the reconstructed tinnitus sound. This suggests to preferably use the reconstructed tinnitus spectrum to determine psychoacoustic tinnitus loudness. Results also support hypothesis (2). Moreover, hearing loss and minimum masking level correlate with oscillatory power in distinctive frequency bands. The lack of an association between psychological comorbidities and oscillatory power may be attributed to the overall low level of mental health problems in the present sample. PMID:23326394
Psychoacoustic cues to emotion in speech prosody and music.
Coutinho, Eduardo; Dibben, Nicola
2013-01-01
There is strong evidence of shared acoustic profiles common to the expression of emotions in music and speech, yet relatively limited understanding of the specific psychoacoustic features involved. This study combined a controlled experiment and computational modelling to investigate the perceptual codes associated with the expression of emotion in the acoustic domain. The empirical stage of the study provided continuous human ratings of emotions perceived in excerpts of film music and natural speech samples. The computational stage created a computer model that retrieves the relevant information from the acoustic stimuli and makes predictions about the emotional expressiveness of speech and music close to the responses of human subjects. We show that a significant part of the listeners' second-by-second reported emotions to music and speech prosody can be predicted from a set of seven psychoacoustic features: loudness, tempo/speech rate, melody/prosody contour, spectral centroid, spectral flux, sharpness, and roughness. The implications of these results are discussed in the context of cross-modal similarities in the communication of emotion in the acoustic domain.
NASA Astrophysics Data System (ADS)
Han, Hyung-Suk
2012-12-01
The indoor noise of a ship is usually determined using the A-weighted sound pressure level. However, in order to better understand this phenomenon, evaluation parameters that more accurately reflect the human sense of hearing are required. To find the level of the satisfaction index of the noise inside a naval vessel such as "Loudness" and "Annoyance", psycho-acoustic evaluation of various sound recordings from the naval vessel was performed in a laboratory. The objective of this paper is to develop a single index of "Loudness" and "Annoyance" for noise inside a naval vessel according to a psycho-acoustic evaluation by using psychological responses such as Noise Rating (NR), Noise Criterion (NC), Room Criterion (RC), Preferred Speech Interference Level (PSIL) and loudness level. Additionally, in order to determine a single index of satisfaction for noise such as "Loudness" and "Annoyance", with respect to a human's sense of hearing, a back-propagation neural network is applied.
Fundamentals of Acoustics. Psychoacoustics and Hearing. Acoustical Measurements
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Ahumada, Al (Technical Monitor)
1997-01-01
These are 3 chapters that will appear in a book titled "Building Acoustical Design", edited by Charles Salter. They are designed to introduce the reader to fundamental concepts of acoustics, particularly as they relate to the built environment. "Fundamentals of Acoustics" reviews basic concepts of sound waveform frequency, pressure, and phase. "Psychoacoustics and Hearing" discusses the human interpretation sound pressure as loudness, particularly as a function of frequency. "Acoustic Measurements" gives a simple overview of the time and frequency weightings for sound pressure measurements that are used in acoustical work.
NASA Astrophysics Data System (ADS)
Vilão, Rui C.; Melo, Santino L. S.
2014-12-01
We address the production of musical tones by a simple musical instrument of the Brazilian tradition: the berimbau-de-barriga. The vibration physics of the string and of the air mass inside the gourd are reviewed. Straightforward measurements of an actual berimbau, which illustrate the basic physical phenomena, are performed using a PC-based "soundcard oscilloscope." The inharmonicity of the string and the role of the gourd are discussed in the context of known results in the psychoacoustics of pitch definition.
Psychoacoustic processing of test signals
NASA Astrophysics Data System (ADS)
Kadlec, Frantisek
2003-10-01
For the quantitative evaluation of electroacoustic system properties and for psychoacoustic testing it is possible to utilize harmonic signals with fixed frequency, sweeping signals, random signals or their combination. This contribution deals with the design of various test signals with emphasis on audible perception. During the digital generation of signals, some additional undesirable frequency components and noise are produced, which are dependent on signal amplitude and sampling frequency. A mathematical analysis describes the origin of this distortion. By proper selection of signal frequency and amplitude it is possible to minimize those undesirable components. An additional step is to minimize the audible perception of this signal distortion by the application of additional noise (dither). For signals intended for listening tests a dither with triangular or Gaussian probability density function was found to be most effective. Signals modified this way may be further improved by the application of noise shaping, which transposes those undesirable products into frequency regions where they are perceived less, according to psychoacoustic principles. The efficiency of individual processing steps was confirmed both by measurements and by listening tests. [Work supported by the Czech Science Foundation.
Initial Investigation into the Psychoacoustic Properties of Small Unmanned Aerial System Noise
NASA Technical Reports Server (NTRS)
Christian, Andrew; Cabell, Randolph
2017-01-01
For the past several years, researchers at NASA Langley have been engaged in a series of projects to study the degree to which existing facilities and capabilities, originally created for work on full-scale aircraft, are extensible to smaller scales --those of the small unmanned aerial systems (sUAS, also UAVs and, colloquially, `drones') that have been showing up in the nation's airspace of late. This paper follows an e ort that has led to an initial human{subject psychoacoustic test regarding the annoyance generated by sUAS noise. This e ort spans three phases: 1. The collection of the sounds through field recordings. 2. The formulation and execution of a psychoacoustic test using those recordings. 3. The initial analysis of the data from that test. The data suggests a lack of parity between the noise of the recorded sUAS and that of a set of road vehicles that were also recorded and included in the test, as measured by a set of contemporary noise metrics. Future work, including the possibility of further human subject testing, is discussed in light of this suggestion.
Efficient FFT Algorithm for Psychoacoustic Model of the MPEG-4 AAC
NASA Astrophysics Data System (ADS)
Lee, Jae-Seong; Lee, Chang-Joon; Park, Young-Cheol; Youn, Dae-Hee
This paper proposes an efficient FFT algorithm for the Psycho-Acoustic Model (PAM) of MPEG-4 AAC. The proposed algorithm synthesizes FFT coefficients using MDCT and MDST coefficients through circular convolution. The complexity of the MDCT and MDST coefficients is approximately half of the original FFT. We also design a new PAM based on the proposed FFT algorithm, which has 15% lower computational complexity than the original PAM without degradation of sound quality. Subjective as well as objective test results are presented to confirm the efficiency of the proposed FFT computation algorithm and the PAM.
Psychoacoustic Assessment to Improve Tinnitus Diagnosis
Hutchins, Sean; Hébert, Sylvie
2013-01-01
The diagnosis of tinnitus relies on self-report. Psychoacoustic measurements of tinnitus pitch and loudness are essential for assessing claims and discriminating true from false ones. For this reason, the quantification of tinnitus remains a challenging research goal. We aimed to: (1) assess the precision of a new tinnitus likeness rating procedure with a continuous-pitch presentation method, controlling for music training, and (2) test whether tinnitus psychoacoustic measurements have the sensitivity and specificity required to detect people faking tinnitus. Musicians and non-musicians with tinnitus, as well as simulated malingerers without tinnitus, were tested. Most were retested several weeks later. Tinnitus pitch matching was first assessed using the likeness rating method: pure tones from 0.25 to 16 kHz were presented randomly to participants, who had to rate the likeness of each tone to their tinnitus, and to adjust its level from 0 to 100 dB SPL. Tinnitus pitch matching was then assessed with a continuous-pitch method: participants had to match the pitch of their tinnitus to an external tone by moving their finger across a touch-sensitive strip, which generated a continuous pure tone from 0.5 to 20 kHz in 1-Hz steps. The predominant tinnitus pitch was consistent across both methods for both musicians and non-musicians, although musicians displayed better external tone pitch matching abilities. Simulated malingerers rated loudness much higher than did the other groups with a high degree of specificity (94.4%) and were unreliable in loudness (not pitch) matching from one session to the other. Retest data showed similar pitch matching responses for both methods for all participants. In conclusion, tinnitus pitch and loudness reliably correspond to the tinnitus percept, and psychoacoustic loudness matches are sensitive and specific to the presence of tinnitus. PMID:24349414
Applications of psychophysical models to the study of auditory development
NASA Astrophysics Data System (ADS)
Werner, Lynne
2003-04-01
Psychophysical models of listening, such as the energy detector model, have provided a framework from which to characterize the function of the mature auditory system and to explore how mature listeners make use of auditory information in sound identification. The application of such models to the study of auditory development has similarly provided insight into the characteristics of infant hearing and listening. Infants intensity, frequency, temporal and spatial resolution have been described at least grossly and some contributions of immature listening strategies to infant hearing have been identified. Infants psychoacoustic performance is typically poorer than adults under identical stimulus conditions. However, the infant's performance typically varies with stimulus condition in a way that is qualitatively similar to the adult's performance. In some cases, though, infants perform in a qualitatively different way from adults in psychoacoustic experiments. Further, recent psychoacoustic studies of children suggest that the classic models of listening may be inadequate to describe the children's performance. The characteristics of a model that might be appropriate for the immature listener will be outlined and the implications for models of mature listening will be discussed. [Work supported by NIH grants DC00396 and by DC04661.
A human-hearing-related prediction tool for soundscapes and community noise
NASA Astrophysics Data System (ADS)
Genuit, Klaus
2002-11-01
There are several methods of calculation available for the prediction of the A-weighted sound-pressure level of environmental noise, which are, however, not suitable for a qualified prediction of the residents' annoyance and physiological strain. The subjectively felt noise quality does not only depend on the A-weighted sound-pressure level, but also on other psychoacoustical parameters, such as loudness, roughness, sharpness, etc. In addition to these physical and psychoacoustical aspects of noise, the so-called psychological or cognitive aspects have to be considered, too, which means that the listeners' expectations, their mental attitude, as well as the information content of the noise finally influence the noise quality perceived by the individual persons. Within the scope of a research project SVEN (Sound Quality of Vehicle Exterior Noise), which is promoted by the EC, a new tool has been developed which allows a binaural simulation and prediction of the environmental noise to evaluate the influence of different contributions by the sound events with respect to the psychoacoustical parameters, the spatial distribution, movement, and frequency. By means of this tool it is now possible to consider completely new aspects regarding the audible perception of noise when establishing a soundscape or when planning community noise.
Biberger, Thomas; Ewert, Stephan D
2017-08-01
The generalized power spectrum model [GPSM; Biberger and Ewert (2016). J. Acoust. Soc. Am. 140, 1023-1038], combining the "classical" concept of the power-spectrum model (PSM) and the envelope power spectrum-model (EPSM), was demonstrated to account for several psychoacoustic and speech intelligibility (SI) experiments. The PSM path of the model uses long-time power signal-to-noise ratios (SNRs), while the EPSM path uses short-time envelope power SNRs. A systematic comparison of existing SI models for several spectro-temporal manipulations of speech maskers and gender combinations of target and masker speakers [Schubotz et al. (2016). J. Acoust. Soc. Am. 140, 524-540] showed the importance of short-time power features. Conversely, Jørgensen et al. [(2013). J. Acoust. Soc. Am. 134, 436-446] demonstrated a higher predictive power of short-time envelope power SNRs than power SNRs using reverberation and spectral subtraction. Here the GPSM was extended to utilize short-time power SNRs and was shown to account for all psychoacoustic and SI data of the three mentioned studies. The best processing strategy was to exclusively use either power or envelope-power SNRs, depending on the experimental task. By analyzing both domains, the suggested model might provide a useful tool for clarifying the contribution of amplitude modulation masking and energetic masking.
Can unaided non-linguistic measures predict cochlear implant candidacy?
Shim, Hyun Joon; Won, Jong Ho; Moon, Il Joon; Anderson, Elizabeth S.; Drennan, Ward R.; McIntosh, Nancy E.; Weaver, Edward M.; Rubinstein, Jay T.
2014-01-01
Objective To determine if unaided, non-linguistic psychoacoustic measures can be effective in evaluating cochlear implant (CI) candidacy. Study Design Prospective split-cohort study including predictor development subgroup and independent predictor validation subgroup. Setting Tertiary referral center. Subjects Fifteen subjects (28 ears) with hearing loss were recruited from patients visiting the University of Washington Medical Center for CI evaluation. Methods Spectral-ripple discrimination (using a 13-dB modulation depth) and temporal modulation detection using 10- and 100-Hz modulation frequencies were assessed with stimuli presented through insert earphones. Correlations between performance for psychoacoustic tasks and speech perception tasks were assessed. Receiver operating characteristic (ROC) curve analysis was performed to estimate the optimal psychoacoustic score for CI candidacy evaluation in the development subgroup and then tested in an independent sample. Results Strong correlations were observed between spectral-ripple thresholds and both aided sentence recognition and unaided word recognition. Weaker relationships were found between temporal modulation detection and speech tests. ROC curve analysis demonstrated that the unaided spectral ripple discrimination shows a good sensitivity, specificity, positive predictive value, and negative predictive value compared to the current gold standard, aided sentence recognition. Conclusions Results demonstrated that the unaided spectral-ripple discrimination test could be a promising tool for evaluating CI candidacy. PMID:24901669
Sanchez, Tanit Ganz; Moraes, Fernanda; Casseb, Juliana; Cota, Jaci; Freire, Katya; Roberts, Larry E.
2016-01-01
Recent neuroscience research suggests that tinnitus may reflect synaptic loss in the cochlea that does not express in the audiogram but leads to neural changes in auditory pathways that reduce sound level tolerance (SLT). Adolescents (N = 170) completed a questionnaire addressing their prior experience with tinnitus, potentially risky listening habits, and sensitivity to ordinary sounds, followed by psychoacoustic measurements in a sound booth. Among all adolescents 54.7% reported by questionnaire that they had previously experienced tinnitus, while 28.8% heard tinnitus in the booth. Psychoacoustic properties of tinnitus measured in the sound booth corresponded with those of chronic adult tinnitus sufferers. Neither hearing thresholds (≤15 dB HL to 16 kHz) nor otoacoustic emissions discriminated between adolescents reporting or not reporting tinnitus in the sound booth, but loudness discomfort levels (a psychoacoustic measure of SLT) did so, averaging 11.3 dB lower in adolescents experiencing tinnitus in the acoustic chamber. Although risky listening habits were near universal, the teenagers experiencing tinnitus and reduced SLT tended to be more protective of their hearing. Tinnitus and reduced SLT could be early indications of a vulnerability to hidden synaptic injury that is prevalent among adolescents and expressed following exposure to high level environmental sounds. PMID:27265722
Ira at 80: The acronyms of a career in acoustics
NASA Astrophysics Data System (ADS)
Weisenberger, Janet M.
2002-05-01
In a career that spans some 54 years to date, the name of Ira J. Hirsh has been associated with significant scientific contributions to psychoacoustics, outstanding mentoring of research scientists, and dedicated service to the fields of acoustics, audiology, and psychology. It is a career that can be traced by acronyms that are part of the daily vocabulary of hearing scientists. These include acronyms of location: Early work at the Psychoacoustics Laboratory at Harvard (PAL), a long tenure in research at the Central Institute for the Deaf (CID), service as faculty member, chair, and dean at Washington University (WashU); acronyms of professional societies that have honored him: Acoustical Society of America (ASA), International Commission of Acoustics (ICA), American Psychological Association (APA), American Psychological Society (APS), American Speech-Language-Hearing Association (ASHA), American Association for the Advancement of Science (AAAS); acronyms of his service to the National Academy of Science: National Research Council (NRC), Commission on Behavioral and Social Science and Education (CBASSE); and acronyms of his contributions to psychoacoustics: Masking Level Difference (MLD), Temporal Order Judgments (TOJ). In large part, these acronyms are part of our vocabulary because of Ira's contributions, and tracing them over the past half-century yields a substantive look at the development of the field of hearing science.
NASA Astrophysics Data System (ADS)
Allen, Prudence
2003-04-01
Young children typically perform more poorly on psychoacoustic tasks than do adults, with large individual differences. When performance is averaged across children within age groups, the data suggest a gradual change in performance with increasing age. However, an examination of individual data suggests that the performance matures more rapidly, although at different times for different children. The mechanisms of development responsible for these changes are likely very complex, involving both sensory and cognitive processes. This paper will discuss some previously suggested mechanisms including attention and cue weighting, as well as possibilities suggested from more recent studies in which learning effects were examined. In one task, a simple frequency discrimination was required, while in another the listener was required to extract regularities in complex sequences of sounds that varied from trial to trial. Results suggested that the ability to select and consistently employ an effective listening strategy was especially important in the performance of the more complex task, while simple stimulus exposure and motivation contributed to the simpler task. These factors are important for understanding the perceptual development and for the subsequent application of psychoacoustic findings to clinical populations. [Work supported by the NSERC and the Canadian Language and Literacy Research Network.
Perceptual-center modeling is affected by including acoustic rate-of-change modulations.
Harsin, C A
1997-02-01
This study investigated the acoustic correlates of perceptual centers (p-centers) in CV and VC syllables and developed an acoustic p-center model. In Part 1, listeners located syllables' p-centers by a method-of-adjustment procedure. The CV syllables contained the consonants /s/,/r/,/n/,/t/,/d/,/k/, and /g/; the VCs, the consonants /s/,/r/, and /n/. The vowel in all syllables was /a/. The results of this experiment replicated and extended previous findings regarding the effects of phonetic variation on p-centers. In Part 2, a digital signal processing procedure was used to acoustically model p-center perception. Each stimulus was passed through a six-band digital filter, and the outputs were processed to derive low-frequency modulation components. These components were weighted according to a perceived modulation magnitude function and recombined to create six psychoacoustic envelopes containing modulation energies from 3 to 47 Hz. In this analysis, p-centers were found to be highly correlated with the time-weighted function of the rate-of-change in the psychoacoustic envelopes, multiplied by the psychoacoustic envelope magnitude increment. The results were interpreted as suggesting (1) the probable role of low-frequency energy modulations in p-center perception, and (2) the presence of perceptual processes that integrate multiple articulatory events into a single syllabic event.
Possibilities of psychoacoustics to determine sound quality
NASA Astrophysics Data System (ADS)
Genuit, Klaus
For some years, acoustic engineers have increasingly become aware of the importance of analyzing and minimizing noise problems not only with regard to the A-weighted sound pressure level, but to design sound quality. It is relatively easy to determine the A-weighted SPL according to international standards. However, the objective evaluation to describe subjectively perceived sound quality - taking into account psychoacoustic parameters such as loudness, roughness, fluctuation strength, sharpness and so forth - is more difficult. On the one hand, the psychoacoustic measurement procedures which are known so far have yet not been standardized. On the other hand, they have only been tested in laboratories by means of listening tests in the free-field and one sound source and simple signals. Therefore, the results achieved cannot be transferred to complex sound situations with several spatially distributed sound sources without difficulty. Due to the directional hearing and selectivity of human hearing, individual sound events can be selected among many. Already in the late seventies a new binaural Artificial Head Measurement System was developed which met the requirements of the automobile industry in terms of measurement technology. The first industrial application of the Artificial Head Measurement System was in 1981. Since that time the system was further developed, particularly by the cooperation between HEAD acoustics and Mercedes-Benz. In addition to a calibratable Artificial Head Measurement System which is compatible with standard measurement technologies and has transfer characteristics comparable to human hearing, a Binaural Analysis System is now also available. This system permits the analysis of binaural signals regarding physical and psychoacoustic procedures. Moreover, the signals to be analyzed can be simultaneously monitored through headphones and manipulated in the time and frequency domain so that those signal components being responsible for noise annoyance can be found. Especially in complex sound situations with several spatially distributed sound sources, standard, one-channel measurements methods cannot adequately determine sound quality, the acoustic comfort, or annoyance of sound events.
NASA Technical Reports Server (NTRS)
Molino, J. A.
1982-01-01
A review of 34 studies indicates that several factors or variables might be important in providing a psychoacoustic foundation for measurements of the noise from helicopters. These factors are phase relations, tail rotor noise, repetition rate, crest level, and generic differences between conventional aircraft and helicopters. Particular attention was given to the impulsive noise known as blade slap. Analysis of the evidence for and against each factor reveals that, for the present state of scientific knowledge, none of these factors should be regarded as the basis for a significant noise measurement correction due to impulsive blade slap. The current method of measuring effective perceived noise level for conventional aircraft appears to be adequate for measuring helicopter noise as well.
Envelope and intensity based prediction of psychoacoustic masking and speech intelligibility.
Biberger, Thomas; Ewert, Stephan D
2016-08-01
Human auditory perception and speech intelligibility have been successfully described based on the two concepts of spectral masking and amplitude modulation (AM) masking. The power-spectrum model (PSM) [Patterson and Moore (1986). Frequency Selectivity in Hearing, pp. 123-177] accounts for effects of spectral masking and critical bandwidth, while the envelope power-spectrum model (EPSM) [Ewert and Dau (2000). J. Acoust. Soc. Am. 108, 1181-1196] has been successfully applied to AM masking and discrimination. Both models extract the long-term (envelope) power to calculate signal-to-noise ratios (SNR). Recently, the EPSM has been applied to speech intelligibility (SI) considering the short-term envelope SNR on various time scales (multi-resolution speech-based envelope power-spectrum model; mr-sEPSM) to account for SI in fluctuating noise [Jørgensen, Ewert, and Dau (2013). J. Acoust. Soc. Am. 134, 436-446]. Here, a generalized auditory model is suggested combining the classical PSM and the mr-sEPSM to jointly account for psychoacoustics and speech intelligibility. The model was extended to consider the local AM depth in conditions with slowly varying signal levels, and the relative role of long-term and short-term SNR was assessed. The suggested generalized power-spectrum model is shown to account for a large variety of psychoacoustic data and to predict speech intelligibility in various types of background noise.
The psychoacoustics of musical articulation
NASA Astrophysics Data System (ADS)
Spiegelberg, Scott Charles
This dissertation develops psychoacoustical definitions of notated articulations, the necessary first step in articulation research. This research can be useful to theorists interested in timbre analysis, the psychology of performance, analysis and performance, the psychology of style differentiation, and performance pedagogy. An explanation of wavelet transforms precedes the development of new techniques for analyzing transient sounds. A history of timbre perception research reveals the inadequacies of current sound segmentation models, resulting in the creation of a new model, the Pitch/Amplitude/Centroid Trajectory (PACT) model of sound segmentation. The new analysis techniques and PACT model are used to analyze recordings of performers playing a melodic fragment in a series of notated articulations. Statistical tests showed that the performers generally agreed on the interpretation of five different articulation groups. A cognitive test of articulation similarity, using musicians and non-musicians as participants, revealed a close correlation between similarity judgments and physical attributes, though additional unknown factors are clearly present. A second psychological test explored the perceptual salience of articulation notation, by asking musically-trained participants to match stimuli to the same notations the performers used. The participants also marked verbal descriptors for each articulation, such as short/long, sharp/dull, loud/soft, harsh/gentle, and normal/extreme. These results were matched against the results of Chapters Five and Six, providing an overall interpretation of the psychoacoustics of articulation.
A Psychoacoustic Evaluation of Noise Signatures from Advanced Civil Transport Aircraft
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.; Christian, Andrew
2016-01-01
The NASA Environmentally Responsible Aviation project has been successful in developing and demonstrating technologies for integrated aircraft systems that can simultaneously meet aggressive goals for fuel burn, noise and emissions. Some of the resulting systems substantially differ from the familiar tube and wing designs constituting the current civil transport fleet. This study attempts to explore whether or not the effective perceived noise level metric used in the NASA noise goal accurately reflects human subject response across the range of vehicles considered. Further, it seeks to determine, in a quantitative manner, if the sounds associated with the advanced aircraft are more or less preferable to the reference vehicles beyond any differences revealed by the metric. These explorations are made through psychoacoustic tests in a controlled laboratory environment using simulated stimuli developed from auralizations of selected vehicles based on systems noise assessments.
Psychoacoustic and cognitive aspects of auditory roughness: definitions, models, and applications
NASA Astrophysics Data System (ADS)
Vassilakis, Pantelis N.; Kendall, Roger A.
2010-02-01
The term "auditory roughness" was first introduced in the 19th century to describe the buzzing, rattling auditory sensation accompanying narrow harmonic intervals (i.e. two tones with frequency difference in the range of ~15-150Hz, presented simultaneously). A broader definition and an overview of the psychoacoustic correlates of the auditory roughness sensation, also referred to as sensory dissonance, is followed by an examination of efforts to quantify it over the past one hundred and fifty years and leads to the introduction of a new roughness calculation model and an application that automates spectral and roughness analysis of sound signals. Implementation of spectral and roughness analysis is briefly discussed in the context of two pilot perceptual experiments, designed to assess the relationship among cultural background, music performance practice, and aesthetic attitudes towards the auditory roughness sensation.
Coutinho, Eduardo; Cangelosi, Angelo
2011-08-01
We sustain that the structure of affect elicited by music is largely dependent on dynamic temporal patterns in low-level music structural parameters. In support of this claim, we have previously provided evidence that spatiotemporal dynamics in psychoacoustic features resonate with two psychological dimensions of affect underlying judgments of subjective feelings: arousal and valence. In this article we extend our previous investigations in two aspects. First, we focus on the emotions experienced rather than perceived while listening to music. Second, we evaluate the extent to which peripheral feedback in music can account for the predicted emotional responses, that is, the role of physiological arousal in determining the intensity and valence of musical emotions. Akin to our previous findings, we will show that a significant part of the listeners' reported emotions can be predicted from a set of six psychoacoustic features--loudness, pitch level, pitch contour, tempo, texture, and sharpness. Furthermore, the accuracy of those predictions is improved with the inclusion of physiological cues--skin conductance and heart rate. The interdisciplinary work presented here provides a new methodology to the field of music and emotion research based on the combination of computational and experimental work, which aid the analysis of the emotional responses to music, while offering a platform for the abstract representation of those complex relationships. Future developments may aid specific areas, such as, psychology and music therapy, by providing coherent descriptions of the emotional effects of specific music stimuli. 2011 APA, all rights reserved
An Overview of Auralization and Psychoacoustics
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Null, Cynthia H. (Technical Monitor)
1998-01-01
This viewgraph presentation provides information on a study the goal of which was to determine the auditory threshold for speech reverberation using a specific room model. Procedures and hardware used are detailed as are the participants in the study.
Epp, Bastian; Yasin, Ifat; Verhey, Jesko L
2013-12-01
The audibility of important sounds is often hampered due to the presence of other masking sounds. The present study investigates if a correlate of the audibility of a tone masked by noise is found in late auditory evoked potentials measured from human listeners. The audibility of the target sound at a fixed physical intensity is varied by introducing auditory cues of (i) interaural target signal phase disparity and (ii) coherent masker level fluctuations in different frequency regions. In agreement with previous studies, psychoacoustical experiments showed that both stimulus manipulations result in a masking release (i: binaural masking level difference; ii: comodulation masking release) compared to a condition where those cues are not present. Late auditory evoked potentials (N1, P2) were recorded for the stimuli at a constant masker level, but different signal levels within the same set of listeners who participated in the psychoacoustical experiment. The data indicate differences in N1 and P2 between stimuli with and without interaural phase disparities. However, differences for stimuli with and without coherent masker modulation were only found for P2, i.e., only P2 is sensitive to the increase in audibility, irrespective of the cue that caused the masking release. The amplitude of P2 is consistent with the psychoacoustical finding of an addition of the masking releases when both cues are present. Even though it cannot be concluded where along the auditory pathway the audibility is represented, the P2 component of auditory evoked potentials is a candidate for an objective measure of audibility in the human auditory system. Copyright © 2013 Elsevier B.V. All rights reserved.
Phase synchronization of instrumental music signals
NASA Astrophysics Data System (ADS)
Mukherjee, Sayan; Palit, Sanjay Kumar; Banerjee, Santo; Ariffin, M. R. K.; Bhattacharya, D. K.
2014-06-01
Signal analysis is one of the finest scientific techniques in communication theory. Some quantitative and qualitative measures describe the pattern of a music signal, vary from one to another. Same musical recital, when played by different instrumentalists, generates different types of music patterns. The reason behind various patterns is the psycho-acoustic measures - Dynamics, Timber, Tonality and Rhythm, varies in each time. However, the psycho-acoustic study of the music signals does not reveal any idea about the similarity between the signals. For such cases, study of synchronization of long-term nonlinear dynamics may provide effective results. In this context, phase synchronization (PS) is one of the measures to show synchronization between two non-identical signals. In fact, it is very critical to investigate any other kind of synchronization for experimental condition, because those are completely non identical signals. Also, there exists equivalence between the phases and the distances of the diagonal line in Recurrence plot (RP) of the signals, which is quantifiable by the recurrence quantification measure τ-recurrence rate. This paper considers two nonlinear music signals based on same raga played by two eminent sitar instrumentalists as two non-identical sources. The psycho-acoustic study shows how the Dynamics, Timber, Tonality and Rhythm vary for the two music signals. Then, long term analysis in the form of phase space reconstruction is performed, which reveals the chaotic phase spaces for both the signals. From the RP of both the phase spaces, τ-recurrence rate is calculated. Finally by the correlation of normalized tau-recurrence rate of their 3D phase spaces and the PS of the two music signals has been established. The numerical results well support the analysis.
Maheras, Kathleen J; Gow, Alexander
2013-09-30
To examine psychoacoustics in mice, we have used 2,2,2-tribromoethanol anesthesia in multiple studies. We find this drug is fast-acting and yields consistent results, providing 25-30 min of anesthesia. Our recent studies in binaural hearing prompted development of a regimen to anesthesia time to 1h. We tested a novel cocktail using 2,2,2-tribromoethanol coupled with low dose chloral hydrate to extend the effective anesthesia time. We have established an intraperitoneal dosing regimen for 2,2,2-tribromoethanol-chloral hydrate anesthesia. To measure efficacy of the drug cocktail, we measured auditory brainstem responses (ABRs) at 10 min intervals to determine the effects on hearing thresholds and wave amplitudes and latencies. This novel drug combination increases effective anesthesia to 1h. ABR Wave I amplitudes, but not latencies, are marginally suppressed. Additionally, amplitudes of the centrally derived Waves III and V show significant inter-animal variability that is independent of stimulus intensity. These data argue against the systematic suppression of ABRs by the drug cocktail. Using 2,2,2-tribromoethanol-chloral hydrate combination in psychoacoustic studies has several advantages over other drug cocktails, the most important being preservation of latencies from centrally- and peripherally-derived ABR waves. In addition, hearing thresholds are unchanged and wave amplitudes are not systematically suppressed, although they exhibit greater variability. We demonstrate that 375 mg/kg 2,2,2-tribromoethanol followed after 5 min by 200mg/kg chloral hydrate provides an anesthesia time of 60 min, has negligible effects on ABR wave latencies and thresholds and non-systematic effects on amplitudes. Copyright © 2013 Elsevier B.V. All rights reserved.
Maheras, Kathleen J.; Gow, Alexander
2013-01-01
Background To examine psychoacoustics in mice, we have used 2,2,2-tribromoethanol anesthesia in multiple studies. We find this drug is fast-acting and yields consistent results, providing 30 – 40 min of anesthesia. Our recent studies in binaural hearing prompted development of a regimen to anesthesia time to one hour. We tested a novel cocktail using 2,2,2-tribromoethanol coupled with low dose chloral hydrate to extend the effective anesthesia time. New Method We have established an intraperitoneal dosing regimen for 2,2,2-tribromoethanol-chloral hydrate anesthesia. To measure efficacy of the drug cocktail, we measured auditory brainstem responses (ABRs) at 10 min intervals to determine the effects on hearing thresholds and wave amplitudes and latencies. Results This novel drug combination increases effective anesthesia to one hour. ABR Wave I amplitudes, but not latencies, are marginally suppressed. Additionally, amplitudes of the centrally-derived Waves III and V show significant inter-animal variability that is independent of stimulus intensity. These data argue against the systematic suppression of ABRs by the drug cocktail. Comparison with Existing Methods Using 2,2,2-tribromoethanol-chloral hydrate combination in psychoacoustic studies has several advantages over other drug cocktails, the most important being preservation of latencies from centrally- and peripherally-derived ABR waves. In addition, hearing thresholds are unchanged and wave amplitudes are not systematically suppressed, although they exhibit greater variability. Conclusions We demonstrate that 375 mg/kg 2,2,2-tribromoethanol followed after five min by 200 mg/kg chloral hydrate provides an anesthesia time of 60 min, has negligible effects on ABR wave latencies and thresholds and non-systematic effects on amplitudes. PMID:23856212
NASA Astrophysics Data System (ADS)
Moore, Brian C. J.
Psychoacoustics
NASA Technical Reports Server (NTRS)
Huston, R. J. (Compiler)
1982-01-01
The establishment of a realistic plan for NASA and the U.S. helicopter industry to develop a design-for-noise methodology, including plans for the identification and development of promising noise reduction technology was discussed. Topics included: noise reduction techniques, scaling laws, empirical noise prediction, psychoacoustics, and methods of developing and validing noise prediction methods.
ERIC Educational Resources Information Center
CLARK, LESLIE L., ED.
THE FOLLOWING FOUR RESEARCH REPORTS ON ORIENTATION AND MOBILITY FOR THE BLIND BASED ON AURAL INFORMATION ARE CONTAINED IN THIS BULLETIN--"A PSYCHOACOUSTIC STUDY OF FACTORS AFFECTING HUMAN ECHOLOCATION" BY JOHN R. WELCH, "ORIENTATION BY AURAL CLUES" BY IVO KOHLER, "SONAR SYSTEM OF THE BLIND" BY WINTHROP N. KELLOGG, AND "TRIAL OF AN ACOUSTIC BLIND…
Impaired Timing and Frequency Discrimination in High-Functioning Autism Spectrum Disorders
ERIC Educational Resources Information Center
Bhatara, Anjali; Babikian, Talin; Laugeson, Elizabeth; Tachdjian, Raffi; Sininger, Yvonne S.
2013-01-01
Individuals with autism spectrum disorders (ASD) frequently demonstrate preserved or enhanced frequency perception but impaired timing perception. The present study investigated the processing of spectral and temporal information in 12 adolescents with ASD and 15 age-matched controls. Participants completed two psychoacoustic tasks: one determined…
Psychoacoustics and Passive Sonar Detection
1974-09-09
al., 1956; Veniar, 1958; Swets, Shipley, McKey und Gree^., 1959; Creelman , 1960; Green, 1961; and Gundy, 1961). In general, the experiments first...Acoust. Soc. Am. 3_7. 1064, (1965). 5. CD. Creelman , "Detection of Signals of Uncertain Frequency", J. Acoust. Soc. Am. 32, 805, (1960). 6. J.P
Auditory Masking Patterns in Bottlenose Dolphins from Anthropogenic and Natural Sound Sources
2011-09-30
in Comparative Psychoacoustics, Birkhäuser Verlag Basel, Switzerland. pp. 307-318. Erbe C. (2008) Critical ratios of beluga whales ( Delphinapterus ... leucas ) and masked signal duration. Journal of the Acoustical Society of America 124:2216-2223. Levitt H. (1971) Transformed up-down methods in
Priming in Melody Perception: Tracking Down the Strength of Cognitive Expectations
ERIC Educational Resources Information Center
Marmel, Frederic; Tillmann, Barbara; Delbe, Charles
2010-01-01
The musical priming paradigm has shown facilitated processing for tonally related over less-related targets. However, the congruence between tonal relatedness and the psychoacoustical properties of music challenges cognitive interpretations of the involved processes. Our goal was to show that cognitive expectations (based on listeners' tonal…
ERIC Educational Resources Information Center
Shinn-Cunningham, Barbara
2017-01-01
Purpose: This review provides clinicians with an overview of recent findings relevant to understanding why listeners with normal hearing thresholds (NHTs) sometimes suffer from communication difficulties in noisy settings. Method: The results from neuroscience and psychoacoustics are reviewed. Results: In noisy settings, listeners focus their…
CLINIC-LABORATORY DESIGN BASED ON FUNCTION AND PHILOSOPHY AT PURDUE UNIVERSITY.
ERIC Educational Resources Information Center
HANLEY, T.D.; STEER, M.D.
THIS REPORT DESCRIBES THE DESIGN OF A NEW CLINIC AND LABORATORY FOR SPEECH AND HEARING TO ACCOMMODATE THE THREE BASIC PROGRAMS OF--(1) CLINICAL TRAINING OF UNDERGRADUATE AND GRADUATE STUDENT MAJORS, (2) SERVICES MADE AVAILABLE TO THE SPEECH AND HEARING HANDICAPPED, AND (3) RESEARCH IN SPEECH PATHOLOGY, AUDIOLOGY, PSYCHO-ACOUSTICS, AND…
Grewe, Oliver; Nagel, Frederik; Kopiez, Reinhard; Altenmüller, Eckart
2005-12-01
Music can arouse ecstatic "chill" experiences defined as "goose pimples" and as "shivers down the spine." We recorded chills both via subjects' self-reports and physiological reactions, finding that they do not occur in a reflex-like manner, but as a result of attentive, experienced, and conscious musical enjoyment.
Psychoacoustic Assessment of Speech Communication Systems. The Diagnostic Discrimination Test.
ERIC Educational Resources Information Center
Grether, Craig Blaine
The present report traces the rationale, development and experimental evaluation of the Diagnostic Discrimination Test (DDT). The DDT is a three-choice test of consonant discriminability of the perceptual/acoustic dimensions of consonant phonemes within specific vowel contexts. The DDT was created and developed in an attempt to provide a…
Binaural frequency selectivity in humans.
Verhey, Jesko L; van de Par, Steven
2018-01-23
Several behavioural studies in humans have shown that listening to sounds with two ears that is binaural hearing, provides the human auditory system with extra information on the sound source that is not available when sounds are only perceived through one ear that is monaurally. Binaural processing involves the analysis of phase and level differences between the two ear signals. As monaural cochlea processing (in each ear) precedes the neural stages responsible for binaural processing properties it is reasonable to assume that properties of the cochlea may also be observed in binaural processing. A main characteristic of cochlea processing is its frequency selectivity. In psychoacoustics, there is an ongoing discussion on the frequency selectivity of the binaural auditory system. While some psychoacoustic experiments seem to indicate poorer frequency selectivity of the binaural system than that of the monaural processing others seem to indicate the same frequency selectivity for monaural and binaural processing. This study provides an overview of these seemingly controversial results and the different explanations that were provided to account for the different results. © 2018 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.
Jain, Chandni; Sahoo, Jitesh Prasad
Tinnitus is the perception of a sound without an external source. It can affect auditory perception abilities in individuals with normal hearing sensitivity. The aim of the study was to determine the effect of tinnitus on psychoacoustic abilities in individuals with normal hearing sensitivity. The study was conducted on twenty subjects with tinnitus and twenty subjects without tinnitus. Tinnitus group was again divided into mild and moderate tinnitus based on the tinnitus handicap inventory. Differential limen of intensity, differential limen of frequency, gap detection test, modulation detection thresholds were done through the mlp toolbox in Matlab and speech in noise test was done with the help of Quick SIN in Kannada. RESULTS of the study showed that the clinical group performed poorly in all the tests except for differential limen of intensity. Tinnitus affects aspects of auditory perception like temporal resolution, speech perception in noise and frequency discrimination in individuals with normal hearing. This could be due to subtle changes in the central auditory system which is not reflected in the pure tone audiogram.
The impact of cochlear fine structure on hearing thresholds and DPOAE levels
NASA Astrophysics Data System (ADS)
Lee, Jungmee; Long, Glenis; Talmadge, Carrick L.
2004-05-01
Although otoacoustic emissions (OAE) are used as clinical and research tools, the correlation between OAE behavioral estimates of hearing status is not large. In normal-hearing individuals, the level of OAEs can vary as much as 30 dB when the frequency is changed less than 5%. These pseudoperiodic variations of OAE level with frequency are known as fine structure. Hearing thresholds measured with high-frequency resolution reveals a similar (up to 15 dB) fine structure. We examine the impact of OAE and threshold fine structures on the prediction of auditory thresholds from OAE levels. Distortion product otoacoustic emissions (DPOAEs) were measured with sweeping primary tones. Psychoacoustic detection thresholds were measured using pure tones, sweep tones, FM tones, and narrow-band noise. Sweep DPOAE and narrow-band threshold estimates provide estimates that are less influenced by cochlear fine structure and should lead to a higher correlation between OAE levels and psychoacoustic thresholds. [Research supported by PSC CUNY, NIDCD, National Institute on Disability and Rehabilitation Research in U.S. Department of Education, and The Ministry of Education in Korea.
Segura-Garcia, Jaume; Navarro-Ruiz, Juan Miguel; Perez-Solano, Juan J; Montoya-Belmonte, Jose; Felici-Castell, Santiago; Cobos, Maximo; Torres-Aranda, Ana M
2018-02-26
Sound pleasantness or annoyance perceived in urban soundscapes is a major concern in environmental acoustics. Binaural psychoacoustic parameters are helpful to describe generic acoustic environments, as it is stated within the ISO 12913 framework. In this paper, the application of a Wireless Acoustic Sensor Network (WASN) to evaluate the spatial distribution and the evolution of urban acoustic environments is described. Two experiments are presented using an indoor and an outdoor deployment of a WASN with several nodes using an Internet of Things (IoT) environment to collect audio data and calculate meaningful parameters such as the sound pressure level, binaural loudness and binaural sharpness. A chunk of audio is recorded in each node periodically with a microphone array and the binaural rendering is conducted by exploiting the estimated directional characteristics of the incoming sound by means of DOA estimation. Each node computes the parameters in a different location and sends the values to a cloud-based broker structure that allows spatial statistical analysis through Kriging techniques. A cross-validation analysis is also performed to confirm the usefulness of the proposed system.
ERIC Educational Resources Information Center
Goverts, S. Theo; Huysmans, Elke; Kramer, Sophia E.; de Groot, Annette M. B.; Houtgast, Tammo
2011-01-01
Purpose: Researchers have used the distortion-sensitivity approach in the psychoacoustical domain to investigate the role of auditory processing abilities in speech perception in noise (van Schijndel, Houtgast, & Festen, 2001; Goverts & Houtgast, 2010). In this study, the authors examined the potential applicability of the…
Psychoacoustic and Electrophysiologic Studies of Hearing under Hyperbaric Pressure.
1980-05-01
and diazepam has been proposed as a supplementary drug . The interaction of ketamine and pressure with and without diazepam as premedication was...eye tracking and angular acceleration exercises. Angular accelerations were administered with a Barany chair arrangement utilizing the Contrave -Goerz... interactions than are characteristic of the narcotized state, where a general depression of transynaptic conduction in peripheral (retinal) and central
The Effect of Remote Masking on the Reception of Speech by Young School-Age Children
ERIC Educational Resources Information Center
Youngdahl, Carla L.; Healy, Eric W.; Yoho, Sarah E.; Apoux, Frédéric; Holt, Rachael Frush
2018-01-01
Purpose: Psychoacoustic data indicate that infants and children are less likely than adults to focus on a spectral region containing an anticipated signal and are more susceptible to remote masking of a signal. These detection tasks suggest that infants and children, unlike adults, do not listen selectively. However, less is known about children's…
Software for Acoustic Rendering
NASA Technical Reports Server (NTRS)
Miller, Joel D.
2003-01-01
SLAB is a software system that can be run on a personal computer to simulate an acoustic environment in real time. SLAB was developed to enable computational experimentation in which one can exert low-level control over a variety of signal-processing parameters, related to spatialization, for conducting psychoacoustic studies. Among the parameters that can be manipulated are the number and position of reflections, the fidelity (that is, the number of taps in finite-impulse-response filters), the system latency, and the update rate of the filters. Another goal in the development of SLAB was to provide an inexpensive means of dynamic synthesis of virtual audio over headphones, without need for special-purpose signal-processing hardware. SLAB has a modular, object-oriented design that affords the flexibility and extensibility needed to accommodate a variety of computational experiments and signal-flow structures. SLAB s spatial renderer has a fixed signal-flow architecture corresponding to a set of parallel signal paths from each source to a listener. This fixed architecture can be regarded as a compromise that optimizes efficiency at the expense of complete flexibility. Such a compromise is necessary, given the design goal of enabling computational psychoacoustic experimentation on inexpensive personal computers.
Loudness of steady sounds - A new theory
NASA Technical Reports Server (NTRS)
Howes, W. L.
1979-01-01
A new mathematical theory for calculating the loudness of steady sounds from power summation and frequency interaction, based on psychoacoustic and physiological information, assuems that loudness is a subjective measure of the electrical energy transmitted along the auditory nerve to the central nervous system. The auditory system consists of the mechanical part modeled by a bandpass filter with a transfer function dependent on the sound pressure, and the electrical part where the signal is transformed into a half-wave reproduction represented by the electrical power in impulsive discharges transmitted along neurons comprising the auditory nerve. In the electrical part the neurons are distributed among artificial parallel channels with frequency bandwidths equal to 'critical bandwidths for loudness', within which loudness is constant for constant sound pressure. The total energy transmitted to the central nervous system is the sum of the energy transmitted in all channels, and the loudness is proportional to the square root of the total filtered sound energy distributed over all channels. The theory explains many psychoacoustic phenomena such as audible beats resulting from closely spaced tones, interaction of sound stimuli which affect the same neurons affecting loudness, and of individually subliminal sounds becoming audible if they lie within the same critical band.
NASA Astrophysics Data System (ADS)
Fishman, Yonatan I.; Arezzo, Joseph C.; Steinschneider, Mitchell
2004-09-01
Auditory stream segregation refers to the organization of sequential sounds into ``perceptual streams'' reflecting individual environmental sound sources. In the present study, sequences of alternating high and low tones, ``...ABAB...,'' similar to those used in psychoacoustic experiments on stream segregation, were presented to awake monkeys while neural activity was recorded in primary auditory cortex (A1). Tone frequency separation (ΔF), tone presentation rate (PR), and tone duration (TD) were systematically varied to examine whether neural responses correlate with effects of these variables on perceptual stream segregation. ``A'' tones were fixed at the best frequency of the recording site, while ``B'' tones were displaced in frequency from ``A'' tones by an amount=ΔF. As PR increased, ``B'' tone responses decreased in amplitude to a greater extent than ``A'' tone responses, yielding neural response patterns dominated by ``A'' tone responses occurring at half the alternation rate. Increasing TD facilitated the differential attenuation of ``B'' tone responses. These findings parallel psychoacoustic data and suggest a physiological model of stream segregation whereby increasing ΔF, PR, or TD enhances spatial differentiation of ``A'' tone and ``B'' tone responses along the tonotopic map in A1.
Transfer of knowledge from sound quality measurement to noise impact evaluation
NASA Astrophysics Data System (ADS)
Genuit, Klaus
2004-05-01
It is well known that the measurement and analysis of sound quality requires a complex procedure with consideration of the physical, psychoacoustical and psychological aspects of sound. Sound quality cannot be described only by a simple value based on A-weighted sound pressure level measurements. The A-weighted sound pressure level is sufficient to predict the probabilty that the human ear could be damaged by sound but the A-weighted level is not the correct descriptor for the annoyance of a complex sound situation given by several different sound events at different and especially moving positions (soundscape). On the one side, the consideration of the spectral distribution and the temporal pattern (psychoacoustics) is requested and, on the other side, the subjective attitude with respect to the sound situation, the expectation and experience of the people (psychology) have to be included in context with the complete noise impact evaluation. This paper describes applications of the newest methods of sound quality measurements-as it is well introduced at the car manufacturers-based on artifical head recordings and signal processing comparable to the human hearing used in noisy environments like community/traffic noise.
Mild Dissonance Preferred Over Consonance in Single Chord Perception
Eerola, Tuomas
2016-01-01
Previous research on harmony perception has mainly been concerned with horizontal aspects of harmony, turning less attention to how listeners perceive psychoacoustic qualities and emotions in single isolated chords. A recent study found mild dissonances to be more preferred than consonances in single chord perception, although the authors did not systematically vary register and consonance in their study; these omissions were explored here. An online empirical experiment was conducted where participants (N = 410) evaluated chords on the dimensions of Valence, Tension, Energy, Consonance, and Preference; 15 different chords were played with piano timbre across two octaves. The results suggest significant differences on all dimensions across chord types, and a strong correlation between perceived dissonance and tension. The register and inversions contributed to the evaluations significantly, nonmusicians distinguishing between triadic inversions similarly to musicians. The mildly dissonant minor ninth, major ninth, and minor seventh chords were rated highest for preference, regardless of musical sophistication. The role of theoretical explanations such as aggregate dyadic consonance, the inverted-U hypothesis, and psychoacoustic roughness, harmonicity, and sharpness will be discussed to account for the preference of mild dissonance over consonance in single chord perception. PMID:27433333
Psychoacoustic influences of the echoing environments of prehistoric art
NASA Astrophysics Data System (ADS)
Waller, Steven J.
2002-11-01
Cave paintings and ancient petroglyphs around the world are typically found in echo rich locations such as caves, canyons, and rocky cliff faces. Analysis of field data shows that echo decibel levels at a large number of prehistoric art sites are higher than those at nondecorated locations. The selection of these echoing environments by the artists appears not to be a mere coincidence. This paper considers the perception of an echoed sound as a psychoacoustic event that would have been inexplicable to ancient humans. A variety of ancient legends from cultures on several continents attribute the phenomenon of echoes to supernatural beings. These legends, together with the quantitative data, strongly implicate echoing as relevant to the artists of the past. The notion that the echoes were caused by spirits within the rock would explain not only the unusual locations of prehistoric art, but also the perplexing subject matter. For example, the common theme of hoofed animal imagery could have been inspired by echoes of percussion noises perceived as hoof beats. Further systematic acoustical studies of prehistoric art sites is warranted. Conservation of the natural acoustic properties of rock art environments--a previously unrecognized need--is urged.
Montoya-Belmonte, Jose; Cobos, Maximo; Torres-Aranda, Ana M.
2018-01-01
Sound pleasantness or annoyance perceived in urban soundscapes is a major concern in environmental acoustics. Binaural psychoacoustic parameters are helpful to describe generic acoustic environments, as it is stated within the ISO 12913 framework. In this paper, the application of a Wireless Acoustic Sensor Network (WASN) to evaluate the spatial distribution and the evolution of urban acoustic environments is described. Two experiments are presented using an indoor and an outdoor deployment of a WASN with several nodes using an Internet of Things (IoT) environment to collect audio data and calculate meaningful parameters such as the sound pressure level, binaural loudness and binaural sharpness. A chunk of audio is recorded in each node periodically with a microphone array and the binaural rendering is conducted by exploiting the estimated directional characteristics of the incoming sound by means of DOA estimation. Each node computes the parameters in a different location and sends the values to a cloud-based broker structure that allows spatial statistical analysis through Kriging techniques. A cross-validation analysis is also performed to confirm the usefulness of the proposed system. PMID:29495407
Teaching hearing science to undergraduate nonscientists
NASA Astrophysics Data System (ADS)
Weiler, Ernest M.; Boyce, Suzanne; Steger, Joseph
2003-04-01
For those students interested in potential clinical careers in Speech Pathology, or Audiology, a knowledge of some of the scientific bases is important, but should not create a distaste for science. The authors have addressed themselves to these goals: (1) calculation of period, Hz, summation of two sine waves, phase and dB; (2) anticipating undergraduate Speech Science; (3) simple examples of hearing pathology; and (4) basic psycho-acoustical issues. The classic material of Harry Helson was used to elucidate issues of context in experimental science, and that of S.S. Stevens was used to exemplify psycho-acoustical formulas of common use. Four texts that have been tried on approximately 200 students were evaluated. Surprisingly, the best provided the fewest formulas, short study questions with answers, good examples, and a list of common terms. The next best was aimed at slightly more advanced students, but each chapter contained introductory material, examples, and definitions suitable for naïve undergraduates. The least satisfactory text provided excerpts of technical material with abrupt transitions, no examples, and only part of the definitions needed for the naïve student. Perhaps the most difficult teaching issue is to avoid demanding graduate-level science from those undergraduates with clinical aspirations.
Auralization of NASA N+2 Aircraft Concepts from System Noise Predictions
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.; Burley, Casey L.; Thomas, Russel H.
2016-01-01
Auralization of aircraft flyover noise provides an auditory experience that complements integrated metrics obtained from system noise predictions. Recent efforts have focused on auralization methods development, specifically the process by which source noise information obtained from semi-empirical models, computational aeroacoustic analyses, and wind tunnel and flight test data, are used for simulated flyover noise at a receiver on the ground. The primary focus of this work, however, is to develop full vehicle auralizations in order to explore the distinguishing features of NASA's N+2 aircraft vis-à-vis current fleet reference vehicles for single-aisle and large twin-aisle classes. Some features can be seen in metric time histories associated with aircraft noise certification, e.g., tone-corrected perceived noise level used in the calculation of effective perceived noise level. Other features can be observed in sound quality metrics, e.g., loudness, sharpness, roughness, fluctuation strength and tone-to-noise ratio. A psychoacoustic annoyance model is employed to establish the relationship between sound quality metrics and noise certification metrics. Finally, the auralizations will serve as the basis for a separate psychoacoustic study aimed at assessing how well aircraft noise certification metrics predict human annoyance for these advanced vehicle concepts.
NASA Astrophysics Data System (ADS)
Cook, Perry R.
This chapter covers algorithms, technologies, computer languages, and systems for computer music. Computer music involves the application of computers and other digital/electronic technologies to music composition, performance, theory, history, and the study of perception. The field combines digital signal processing, computational algorithms, computer languages, hardware and software systems, acoustics, psychoacoustics (low-level perception of sounds from the raw acoustic signal), and music cognition (higher-level perception of musical style, form, emotion, etc.).
Challenges to the successful implementation of 3-D sound
NASA Astrophysics Data System (ADS)
Begault, Durand R.
1991-11-01
The major challenges for the successful implementation of 3-D audio systems involve minimizing reversals, intracranially heard sound, and localization error for listeners. Designers of 3-D audio systems are faced with additional challenges in data reduction and low-frequency response characteristics. The relationship of the head-related transfer function (HRTF) to these challenges is shown, along with some preliminary psychoacoustic results gathered at NASA-Ames.
1989-06-01
the intensity for which performance equals the chosen value. We use the PEST (parameter estimation by sequential testing; Taylor and Creelman , 1967...forward masking in the auditory nerve." J. Acoust. Soc. Am. 84, 584-591. Taylor, M.M. and Creelman , C.D. (1967). "PEST: Efficient estimates on
Good vibrations: "sirens," soundscapes, and physiology.
Plock, Vike Martina
2008-01-01
This article establishes Joyce's ongoing interest in psychoacoustics and illustrates how much he drew, in the writing of the "Sirens" episode, on nineteenth-century sound experiments that were developed by the German physician Hermann von Helmholtz. It argues that Joyce consciously referenced nineteenth-century sound theories to explore the link between the emotional and sensory experience of music and the physical and physiological components of sound perception.
Assessment of Ipsilateral Efferent Effects in Human via ECochG
Verschooten, Eric; Strickland, Elizabeth A.; Verhaert, Nicolas; Joris, Philip X.
2017-01-01
Development of electrophysiological means to assess the medial olivocochlear (MOC) system in humans is important to further our understanding of the function of that system and for the refinement and validation of psychoacoustical and otoacoustic emission methods which are thought to probe the MOC. Based on measurements in anesthetized animals it has been hypothesized that the MOC-reflex (MOCR) can enhance the response to signals in noise, and several lines of evidence support such a role in humans. A difficulty in these studies is the isolation of efferent effects. Efferent activation can be triggered by acoustic stimulation of the contralateral or ipsilateral ear, but ipsilateral stimulation is thought to be more effective. However, ipsilateral stimulation complicates interpretation of effects since these sounds can affect the perception of other ipsilateral sounds by mechanisms not involving olivocochlear efferents. We assessed the ipsilaterally evoked MOCR in human using a transtympanic procedure to record mass-potentials from the cochlear promontory or the niche of the round window. Averaged compound action potential (CAP) responses to masked probe tones of 4 kHz with and without a precursor (designed to activate the MOCR but not the stapedius reflex) were extracted with a polarity alternating paradigm. The masker was either a simultaneous narrow band noise masker or a short (20-ms) tonal ON- or OFF-frequency forward masker. The subjects were screened for normal hearing (audiogram, tympanogram, threshold stapedius reflex) and psychoacoustically tested for the presence of a precursor effect. We observed a clear reduction of CAP amplitude by the precursor, for different masking conditions. Even without an MOCR, this is expected because the precursor will affect the response to subsequent stimuli via neural adaptation. To determine whether the precursor also activated the efferent system, we measured the CAP over a range of masker levels, with or without precursor, and for different types of masker. The results show CAP reduction consistent with the type of gain reduction caused by the MOCR. These results generally support psychoacoustical paradigms designed to probe the efferent system as indeed activating the MOCR system, but not all observations are consistent with this mechanism. PMID:28642679
The neural processing of masked speech
Scott, Sophie K; McGettigan, Carolyn
2014-01-01
Spoken language is rarely heard in silence, and a great deal of interest in psychoacoustics has focused on the ways that the perception of speech is affected by properties of masking noise. In this review we first briefly outline the neuroanatomy of speech perception. We then summarise the neurobiological aspects of the perception of masked speech, and investigate this as a function of masker type, masker level and task. PMID:23685149
Włodarczyk, Elżbieta; Szkiełkowska, Agata; Skarżyński, Henryk; Piłka, Adam
2011-01-01
To assess effectiveness of the auditory training in children with dyslalia and central auditory processing disorders. Material consisted of 50 children aged 7-9-years-old. Children with articulation disorders stayed under long-term speech therapy care in the Auditory and Phoniatrics Clinic. All children were examined by a laryngologist and a phoniatrician. Assessment included tonal and impedance audiometry and speech therapists' and psychologist's consultations. Additionally, a set of electrophysiological examinations was performed - registration of N2, P2, N2, P2, P300 waves and psychoacoustic test of central auditory functions: FPT - frequency pattern test. Next children took part in the regular auditory training and attended speech therapy. Speech assessment followed treatment and therapy, again psychoacoustic tests were performed and P300 cortical potentials were recorded. After that statistical analyses were performed. Analyses revealed that application of auditory training in patients with dyslalia and other central auditory disorders is very efficient. Auditory training may be a very efficient therapy supporting speech therapy in children suffering from dyslalia coexisting with articulation and central auditory disorders and in children with educational problems of audiogenic origin. Copyright © 2011 Polish Otolaryngology Society. Published by Elsevier Urban & Partner (Poland). All rights reserved.
The Effect of Remote Masking on the Reception of Speech by Young School-Age Children.
Youngdahl, Carla L; Healy, Eric W; Yoho, Sarah E; Apoux, Frédéric; Holt, Rachael Frush
2018-02-15
Psychoacoustic data indicate that infants and children are less likely than adults to focus on a spectral region containing an anticipated signal and are more susceptible to remote masking of a signal. These detection tasks suggest that infants and children, unlike adults, do not listen selectively. However, less is known about children's ability to listen selectively during speech recognition. Accordingly, the current study examines remote masking during speech recognition in children and adults. Adults and 7- and 5-year-old children performed sentence recognition in the presence of various spectrally remote maskers. Intelligibility was determined for each remote-masker condition, and performance was compared across age groups. It was found that speech recognition for 5-year-olds was reduced in the presence of spectrally remote noise, whereas the maskers had no effect on the 7-year-olds or adults. Maskers of different bandwidth and remoteness had similar effects. In accord with psychoacoustic data, young children do not appear to focus on a spectral region of interest and ignore other regions during speech recognition. This tendency may help account for their typically poorer speech perception in noise. This study also appears to capture an important developmental stage, during which a substantial refinement in spectral listening occurs.
Yang, Ming; De Coensel, Bert; Kang, Jian
2015-08-01
1/f noise or pink noise, which has been shown to be universal in nature, has also been observed in the temporal envelope of music, speech, and environmental sound. Moreover, the slope of the spectral density of the temporal envelope of music has been shown to correlate well to its pleasing, dull, or chaotic character. In this paper, the temporal structure of a number of instantaneous psychoacoustic parameters of environmental sound is examined in order to investigate whether a 1/f temporal structure appears in various types of sound that are generally preferred by people in everyday life. The results show, to some extent, that different categories of environmental sounds have different temporal structure characteristics. Only a number of urban sounds considered and birdsong, generally, exhibit 1/f behavior on short to medium duration time scales, i.e., from 0.1 s to 10 s, in instantaneous loudness and sharpness, whereas a more chaotic variation is found in birdsong at longer time scales, i.e., of 10 s-200 s. The other sound categories considered exhibit random or monotonic variations in the different time scales. In general, this study shows that a 1/f temporal structure is not necessarily present in environmental sounds that are commonly perceived as pleasant.
Psychoacoustic Analysis of Synthesized Jet Noise
NASA Technical Reports Server (NTRS)
Okcu, Selen; Rathsam, Jonathan; Rizzi, Stephen A.
2013-01-01
An aircraft noise synthesis capability is being developed so the annoyance caused by proposed aircraft can be assessed during the design stage. To make synthesized signals as realistic as possible, high fidelity simulation is required for source (e.g., engine noise, airframe noise), propagation and receiver effects. This psychoacoustic study tests whether the jet noise component of synthesized aircraft engine noise can be made more realistic using a low frequency oscillator (LFO) technique to simulate fluctuations in level observed in recordings. Jet noise predictions are commonly made in the frequency domain based on models of time-averaged empirical data. The synthesis process involves conversion of the frequency domain prediction into an audible pressure time history. However, because the predictions are time-invariant, the synthesized sound lacks fluctuations observed in recordings. Such fluctuations are hypothesized to be perceptually important. To introduce time-varying characteristics into jet noise synthesis, a method has been developed that modulates measured or predicted 1/3-octave band levels with a (<20Hz) LFO. The LFO characteristics are determined through analysis of laboratory jet noise recordings. For the aft emission angle, results indicate that signals synthesized using a generic LFO are perceived as more similar to recordings than those using no LFO, and signals synthesized with an angle-specific LFO are more similar to recordings than those synthesized with a generic LFO.
Psychoacoustic Testing of Modulated Blade Spacing for Main Rotors
NASA Technical Reports Server (NTRS)
Edwards, Bryan; Booth, Earl R., Jr. (Technical Monitor)
2002-01-01
Psychoacoustic testing of simulated helicopter main rotor noise is described, and the subjective results are presented. The objective of these tests was to evaluate the potential acoustic benefits of main rotors with modulated (uneven) blade spacing. Sound simulations were prepared for six main rotor configurations. A baseline 4-blade main rotor with regular blade spacing was based on the Bell Model 427 helicopter. A 5-blade main rotor with regular spacing was designed to approximate the performance of the 427, but at reduced tipspeed. Four modulated rotors - one with "optimum" spacing and three alternate configurations - were derived from the 5 bladed regular spacing rotor. The sounds were played to 2 subjects at a time, with care being taken in the speaker selection and placement to ensure that the sounds were identical for each subject. A total of 40 subjects participated. For each rotor configuration, the listeners were asked to evaluate the sounds in terms of noisiness. The test results indicate little to no "annoyance" benefit for the modulated blade spacing. In general, the subjects preferred the sound of the 5-blade regular spaced rotor over any of the modulated ones. A conclusion is that modulated blade spacing is not a promising design feature to reduce the annoyance for helicopter main rotors.
Díaz, Begoña; Baus, Cristina; Escera, Carles; Costa, Albert; Sebastián-Gallés, Núria
2008-01-01
Human beings differ in their ability to master the sounds of their second language (L2). Phonetic training studies have proposed that differences in phonetic learning stem from differences in psychoacoustic abilities rather than speech-specific capabilities. We aimed at finding the origin of individual differences in L2 phonetic acquisition in natural learning contexts. We consider two alternative explanations: a general psychoacoustic origin vs. a speech-specific one. For this purpose, event-related potentials (ERPs) were recorded from two groups of early, proficient Spanish-Catalan bilinguals who differed in their mastery of the Catalan (L2) phonetic contrast /e-ε/. Brain activity in response to acoustic change detection was recorded in three different conditions involving tones of different length (duration condition), frequency (frequency condition), and presentation order (pattern condition). In addition, neural correlates of speech change detection were also assessed for both native (/o/-/e/) and nonnative (/o/-/ö/) phonetic contrasts (speech condition). Participants' discrimination accuracy, reflected electrically as a mismatch negativity (MMN), was similar between the two groups of participants in the three acoustic conditions. Conversely, the MMN was reduced in poor perceivers (PP) when they were presented with speech sounds. Therefore, our results support a speech-specific origin of individual variability in L2 phonetic mastery. PMID:18852470
Signal Feature Analysis Using Neural Networks & Psychoacoustics
1993-05-01
large class file on the DAT recording . This processing produced signals which ranged in length from 13200 and 39650 points. The extractions produced ... recorded . This signal set, denoted as "Air" signals , lacked the parameter of angle but added the parameter of striker (metal, plastic, and wood...the subjects were recorded . These became r.4 w data for confusion matrices which described how often a subject confused the class of a signal
Low Frequency Acoustic Detection Research in Support of Human Detection Range Prediction
1979-10-01
beat at narrow separations and hence made estimates of bandwidth difficult. In addition, Zwicker’s and Green’s data show large discrepancies, the...already known that this spurious low frequency noise can profoundly influence psychoacoustic results. For some years a binaural phenomenon known as the...tend to be uncorrelated in the two ears) and thus preserved the binaural advantage for the low frequency signals. Green et al. (Reference 21) used a
Research and Development for Digital Voice Processing.
1986-10-01
9.6 kbps. Analysis of the DRT data shows that the psychoacousticly based "channel vocoder" has difficulty with the speech attributes sustention and...thought as fought or did a-b--d (graveness). Sustention -absent or "interrupted" correlates to an abrupt onset of energy across the the full spectrum with...system, Table 4.1 presents the board number(s), a description, memory address(es), interrupt vectcr location(s), and bus priority. Note that several
Patterns of phonological disability in Cantonese-speaking children in Hong Kong.
Cheung, P; Abberton, E
2000-01-01
Tone, vowel and consonant production are described for a large group of Cantonese-speaking children assessed in speech and language therapy clinics in Hong Kong. The patterns of disability follow predictions made on the basis of work on normal phonological development in Cantonese, and on psychoacoustic factors in acquisition: consonants account for more disability than vowels, and tones are least problematic. Possible articulatory and auditory contributions to explanation of the observed patterns are discussed.
Sonic Boom Modeling Technical Challenge
NASA Technical Reports Server (NTRS)
Sullivan, Brenda M.
2007-01-01
This viewgraph presentation reviews the technical challenges in modeling sonic booms. The goal of this program is to develop knowledge, capabilities and technologies to enable overland supersonic flight. The specific objectives of the modeling are: (1) Develop and validate sonic boom propagation model through realistic atmospheres, including effects of turbulence (2) Develop methods enabling prediction of response of and acoustic transmission into structures impacted by sonic booms (3) Develop and validate psychoacoustic model of human response to sonic booms under both indoor and outdoor listening conditions, using simulators.
Application of subharmonics for active sound design of electric vehicles.
Gwak, Doo Young; Yoon, Kiseop; Seong, Yeolwan; Lee, Soogab
2014-12-01
The powertrain of electric vehicles generates an unfamiliar acoustical environment for customers. This paper seeks optimal interior sound for electric vehicles based on psychoacoustic knowledge and musical harmonic theory. The concept of inserting a virtual sound, which consists of the subharmonics of an existing high-frequency component, is suggested to improve sound quality. Subjective evaluation results indicate that the impression of interior sound can be enhanced in this manner. Increased appeal is achieved through two designed stimuli, which proves the effectiveness of the method proposed.
Effect of Human Auditory Efferent Feedback on Cochlear Gain and Compression
Drga, Vit; Plack, Christopher J.
2014-01-01
The mammalian auditory system includes a brainstem-mediated efferent pathway from the superior olivary complex by way of the medial olivocochlear system, which reduces the cochlear response to sound (Warr and Guinan, 1979; Liberman et al., 1996). The human medial olivocochlear response has an onset delay of between 25 and 40 ms and rise and decay constants in the region of 280 and 160 ms, respectively (Backus and Guinan, 2006). Physiological studies with nonhuman mammals indicate that onset and decay characteristics of efferent activation are dependent on the temporal and level characteristics of the auditory stimulus (Bacon and Smith, 1991; Guinan and Stankovic, 1996). This study uses a novel psychoacoustical masking technique using a precursor sound to obtain a measure of the efferent effect in humans. This technique avoids confounds currently associated with other psychoacoustical measures. Both temporal and level dependency of the efferent effect was measured, providing a comprehensive measure of the effect of human auditory efferents on cochlear gain and compression. Results indicate that a precursor (>20 dB SPL) induced efferent activation, resulting in a decrease in both maximum gain and maximum compression, with linearization of the compressive function for input sound levels between 50 and 70 dB SPL. Estimated gain decreased as precursor level increased, and increased as the silent interval between the precursor and combined masker-signal stimulus increased, consistent with a decay of the efferent effect. Human auditory efferent activation linearizes the cochlear response for mid-level sounds while reducing maximum gain. PMID:25392499
Bigand, Emmanuel; Delbé, Charles; Poulin-Charronnat, Bénédicte; Leman, Marc; Tillmann, Barbara
2014-01-01
During the last decade, it has been argued that (1) music processing involves syntactic representations similar to those observed in language, and (2) that music and language share similar syntactic-like processes and neural resources. This claim is important for understanding the origin of music and language abilities and, furthermore, it has clinical implications. The Western musical system, however, is rooted in psychoacoustic properties of sound, and this is not the case for linguistic syntax. Accordingly, musical syntax processing could be parsimoniously understood as an emergent property of auditory memory rather than a property of abstract processing similar to linguistic processing. To support this view, we simulated numerous empirical studies that investigated the processing of harmonic structures, using a model based on the accumulation of sensory information in auditory memory. The simulations revealed that most of the musical syntax manipulations used with behavioral and neurophysiological methods as well as with developmental and cross-cultural approaches can be accounted for by the auditory memory model. This led us to question whether current research on musical syntax can really be compared with linguistic processing. Our simulation also raises methodological and theoretical challenges to study musical syntax while disentangling the confounded low-level sensory influences. In order to investigate syntactic abilities in music comparable to language, research should preferentially use musical material with structures that circumvent the tonal effect exerted by psychoacoustic properties of sounds. PMID:24936174
Role of worry in patients with chronic tinnitus and sensorineural hearing loss: a preliminary study.
Caldirola, Daniela; Teggi, Roberto; Daccò, Silvia; Sangiorgio, Erika; Bussi, Mario; Perna, Giampaolo
2016-12-01
Tinnitus-related distress appears to be more strongly associated with multiple psychological factors than with any perceptual properties of tinnitus. Prior studies have not investigated the role of worry in tinnitus sufferers. Worry is a dispositional cognitive trait that involves a pervasive, non-specific, future-oriented proneness to fretting, which can foster negative affective states and catastrophic thinking about a specific trouble when the trouble is actual and present. We examined the relationship between worry and self-perceived anxiety and depressive symptoms and handicap in 54 outpatients with chronic tinnitus and sensorineural hearing loss who had been previously recruited for a randomized double-blind study on the efficacy of transmeatal low-level laser therapy for tinnitus. We measured the current anxiety and depressive symptoms with the State-Trait Anxiety Inventory Form Y-1/Self-evaluation Depression Scale, the handicap with the Tinnitus Handicap Inventory, and the proneness to worry with the Penn State Worry Questionnaire. For the psychoacoustic tinnitus measures, we considered the loudness match and the minimum masking level. We found that tinnitus-related anxiety and depressive symptoms and handicap were significantly associated with proneness to worry (linear regression models, p < 0.01), whereas no associations were found with the psychoacoustic measures. This suggests the usefulness of worry assessment when managing chronic tinnitus in clinical practice. Early therapeutic interventions for reducing proneness to worry may facilitate better adaptation to tinnitus.
An automated psychoacoustic testing apparatus for use in cats.
Benovitski, Yuri B; Blamey, Peter J; Rathbone, Graeme D; Fallon, James B
2014-03-01
Animal behavioral studies make a significant contribution to hearing research and provide vital information which is not available from human subjects. Animal psychoacoustics is usually extremely time consuming and labor intensive; in addition, animals may become stressed, especially if restraints or negative reinforcers such as electric shocks are used. We present a novel behavioral experimental system that was developed to allow efficient animal training in response to acoustic stimuli. Cats were required to perform a relatively simple task of moving toward and away from the device depending on whether the members of a tone pair were different or the same in frequency (go/no-go task). The experimental setup proved to be effective, with all animals (N = 7) performing at above 90% correct on an easy task. Animals were trained within 2-4 weeks and then generated a total of 150-200 trials per day, distributed within approximately 8 self initiated sessions. Data collected using this system were stable over 1 week and repeatable over long test periods (14 weeks). Measured frequency discrimination thresholds from 3 animals at 3 different reference frequencies were comparable with previously published results. The main advantages of the system are: relatively simple setup; large amounts of data can be generated without the need of researcher supervision; multiple animals can be tested simultaneously without removal from home pens; and no electric shocks or restraints are required. Copyright © 2013 Elsevier B.V. All rights reserved.
Function and Evolution of Vibrato-like Frequency Modulation in Mammals.
Charlton, Benjamin D; Taylor, Anna M; Reby, David
2017-09-11
Why do distantly related mammals like sheep, giant pandas, and fur seals produce bleats that are characterized by vibrato-like fundamental frequency (F0) modulation? To answer this question, we used psychoacoustic tests and comparative analyses to investigate whether this distinctive vocal feature has evolved to improve the perception of formants, key acoustic components of animal calls that encode important information about the caller's size and identity [1]. Psychoacoustic tests on humans confirmed that vibrato-like F0 modulation improves the ability of listeners to detect differences in the formant patterns of synthetic bleat-like stimuli. Subsequent phylogenetically controlled comparative analyses revealed that vibrato-like F0 modulation has evolved independently in six mammalian orders in vocal signals with relatively high F0 and, therefore, low spectral density (i.e., less harmonic overtones). We also found that mammals modulate the vibrato in these calls over greater frequency extents when the number of harmonic overtones per formant is low, suggesting that this is a mechanism to improve formant perception in calls with low spectral density. Our findings constitute the first evidence that formant perception in non-speech sounds is improved by fundamental frequency modulation and provide a mechanism for the convergent evolution of bleat-like calls in mammals. They also indicate that selection pressures for animals to transmit important information encoded by formant frequencies (on size and identity, for example) are likely to have been a key driver in the evolution of mammal vocal diversity. Copyright © 2017 Elsevier Ltd. All rights reserved.
Schädler, Marc René; Warzybok, Anna; Ewert, Stephan D; Kollmeier, Birger
2016-05-01
A framework for simulating auditory discrimination experiments, based on an approach from Schädler, Warzybok, Hochmuth, and Kollmeier [(2015). Int. J. Audiol. 54, 100-107] which was originally designed to predict speech recognition thresholds, is extended to also predict psychoacoustic thresholds. The proposed framework is used to assess the suitability of different auditory-inspired feature sets for a range of auditory discrimination experiments that included psychoacoustic as well as speech recognition experiments in noise. The considered experiments were 2 kHz tone-in-broadband-noise simultaneous masking depending on the tone length, spectral masking with simultaneously presented tone signals and narrow-band noise maskers, and German Matrix sentence test reception threshold in stationary and modulated noise. The employed feature sets included spectro-temporal Gabor filter bank features, Mel-frequency cepstral coefficients, logarithmically scaled Mel-spectrograms, and the internal representation of the Perception Model from Dau, Kollmeier, and Kohlrausch [(1997). J. Acoust. Soc. Am. 102(5), 2892-2905]. The proposed framework was successfully employed to simulate all experiments with a common parameter set and obtain objective thresholds with less assumptions compared to traditional modeling approaches. Depending on the feature set, the simulated reference-free thresholds were found to agree with-and hence to predict-empirical data from the literature. Across-frequency processing was found to be crucial to accurately model the lower speech reception threshold in modulated noise conditions than in stationary noise conditions.
Factors Affecting Acoustics and Speech Intelligibility in the Operating Room: Size Matters.
McNeer, Richard R; Bennett, Christopher L; Horn, Danielle Bodzin; Dudaryk, Roman
2017-06-01
Noise in health care settings has increased since 1960 and represents a significant source of dissatisfaction among staff and patients and risk to patient safety. Operating rooms (ORs) in which effective communication is crucial are particularly noisy. Speech intelligibility is impacted by noise, room architecture, and acoustics. For example, sound reverberation time (RT60) increases with room size, which can negatively impact intelligibility, while room objects are hypothesized to have the opposite effect. We explored these relationships by investigating room construction and acoustics of the surgical suites at our institution. We studied our ORs during times of nonuse. Room dimensions were measured to calculate room volumes (VR). Room content was assessed by estimating size and assigning items into 5 volume categories to arrive at an adjusted room content volume (VC) metric. Psychoacoustic analyses were performed by playing sweep tones from a speaker and recording the impulse responses (ie, resulting sound fields) from 3 locations in each room. The recordings were used to calculate 6 psychoacoustic indices of intelligibility. Multiple linear regression was performed using VR and VC as predictor variables and each intelligibility index as an outcome variable. A total of 40 ORs were studied. The surgical suites were characterized by a large degree of construction and surface finish heterogeneity and varied in size from 71.2 to 196.4 m (average VR = 131.1 [34.2] m). An insignificant correlation was observed between VR and VC (Pearson correlation = 0.223, P = .166). Multiple linear regression model fits and β coefficients for VR were highly significant for each of the intelligibility indices and were best for RT60 (R = 0.666, F(2, 37) = 39.9, P < .0001). For Dmax (maximum distance where there is <15% loss of consonant articulation), both VR and VC β coefficients were significant. For RT60 and Dmax, after controlling for VC, partial correlations were 0.825 (P < .0001) and 0.718 (P < .0001), respectively, while after controlling for VR, partial correlations were -0.322 (P = .169) and 0.381 (P < .05), respectively. Our results suggest that the size and contents of an OR can predict a range of psychoacoustic indices of speech intelligibility. Specifically, increasing OR size correlated with worse speech intelligibility, while increasing amounts of OR contents correlated with improved speech intelligibility. This study provides valuable descriptive data and a predictive method for identifying existing ORs that may benefit from acoustic modifiers (eg, sound absorption panels). Additionally, it suggests that room dimensions and projected clinical use should be considered during the design phase of OR suites to optimize acoustic performance.
Bilateral cochlear implantation in a patient with bilateral temporal bone fractures.
Chung, Jae Ho; Shin, Myung Chul; Min, Hyun Jung; Park, Chul Won; Lee, Seung Hwan
2011-01-01
With the emphasis on bilateral hearing nowadays, bilateral cochlear implantation has been tried out for bilateral aural rehabilitation. Bilateral sensorineural hearing loss caused by head trauma can get help from cochlear implantation. We present the case of a 44-year-old man with bilateral otic capsule violating temporal bone fractures due to head trauma. The patient demonstrated much improved audiometric and psychoacoustic performance after bilateral cochlear implantation. We believe bilateral cochlear implantation in such patient can be a very effective tool for rehabilitation. Copyright © 2011 Elsevier Inc. All rights reserved.
USAF bioenvironmental noise data handbook. Volume 148. T-37B in-flight crew noise
NASA Astrophysics Data System (ADS)
Hille, H. K.
1981-11-01
The T-37B is a USAF two-seat primary trainer aircraft. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this aircraft during normal flight operations. Data are reported at one location for 19 different flight conditions and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors.
A Three-Dimensional Virtual Simulator for Aircraft Flyover Presentation
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.; Sullivan, Brenda M.; Sandridge, Christopher A.
2003-01-01
This paper presents a system developed at NASA Langley Research Center to render aircraft flyovers in a virtual reality environment. The present system uses monaural recordings of actual aircraft flyover noise and presents these binaurally using head tracking information. The three-dimensional audio is simultaneously rendered with a visual presentation using a head-mounted display (HMD). The final system will use flyover noise synthesized using data from various analytical and empirical modeling systems. This will permit presentation of flyover noise from candidate low-noise flight operations to subjects for psychoacoustical evaluation.
A new method for measuring the psychoacoustical properties of tinnitus
2013-01-01
Background This study investigates the usefulness and effectiveness of a new way of tinnitus screening and diagnosing. The authors believe that in order to arrive at relevant diagnostic information, select the tinnitus treatment and quantitatively substantiate its effects, the measurement of the Tinnitus psychoacoustic parameters should be made an inherent part of the Tinnitus therapy. Methods For this purpose the multimedia-based sound synthesizer has been proposed for testing tinnitus and the results obtained this way are compared with the outcome of the audiometer-based Wilcoxon test. The method has been verified with 14 patients suffering from tinnitus. Results The experiments reveal capabilities, limitations, advantages and disadvantages of both methods. The synthesizer enables the patient to estimate his/her tinnitus more than twice as fast as the audiometer and makes the information on the tinnitus character perception more accurate. The analysis of the Wilcoxon test results shows that there are statistically important differences between the two tests. Conclusions Patients using the synthesizer operate the software application themselves and thus get more involved in testing. Moreover, they do not concentrate on describing verbally their tinnitus, which could be difficult for some of them. As a result, the test outcome is closer to the perceived tinnitus. However, the more complex the description of the perceived tinnitus, the harder it is to determine the sound parameters of the patient’s perception. It also takes more time regardless of the method. Virtual slides The virtual slide(s) for this article can be found here: http://www.diagnosticpathology.diagnomx.eu/vs/1954066324109436 PMID:24354736
Alves-Pinto, A.; Sollini, J.; Sumner, C.J.
2012-01-01
Signal detection theory (SDT) provides a framework for interpreting psychophysical experiments, separating the putative internal sensory representation and the decision process. SDT was used to analyse ferret behavioural responses in a (yes–no) tone-in-noise detection task. Instead of measuring the receiver-operating characteristic (ROC), we tested SDT by comparing responses collected using two common psychophysical data collection methods. These (Constant Stimuli, Limits) differ in the set of signal levels presented within and across behavioural sessions. The results support the use of SDT as a method of analysis: SDT sensory component was unchanged between the two methods, even though decisions depended on the stimuli presented within a behavioural session. Decision criterion varied trial-by-trial: a ‘yes’ response was more likely after a correct rejection trial than a hit trial. Simulation using an SDT model with several decision components reproduced the experimental observations accurately, leaving only ∼10% of the variance unaccounted for. The model also showed that trial-by-trial dependencies were unlikely to influence measured psychometric functions or thresholds. An additional model component suggested that inattention did not contribute substantially. Further analysis showed that ferrets were changing their decision criteria, almost optimally, to maximise the reward obtained in a session. The data suggest trial-by-trial reward-driven optimization of the decision process. Understanding the factors determining behavioural responses is important for correlating neural activity and behaviour. SDT provides a good account of animal psychoacoustics, and can be validated using standard psychophysical methods and computer simulations, without recourse to ROC measurements. PMID:22698686
Registers in Infant Phonation.
Buder, Eugene H; McDaniel, Valerie F; Bene, Edina R; Ladmirault, Jennifer; Oller, D Kimbrough
2018-04-09
The primary vocal registers of modal, falsetto, and fry have been studied in adults but not per se in infancy. The vocal ligament is thought to play a critical role in the modal-falsetto contrast but is still developing during infancy (Tateya and Tateya, 2015). 41 Cover tissues are also implicated in the modal-fry contrast, but the low fundamental frequency (f o ) cutoff of 70 Hz, shared between genders, suggests a psychoacoustic basis for the contrast. Buder, Chorna, Oller, and Robinson (2008) 6 used the labels of "loft," "modal," and "pulse" for distinct vibratory regimes that appear to be identifiable based on spectrographic inspection of harmonic structure and auditory judgments in infants, but this work did not supply acoustic measurements to verify which of these nominally labeled regimes resembled adult registers. In this report, we identify clear transitions between registers within infant vocalizations and measure these registers and their transitions for f o and relative harmonic amplitudes (H1-H2). By selectively sampling first-year vocalizations, this manuscript quantifies acoustic patterns that correspond to vocal fold vibration types not previously cataloged in infancy. Results support a developmental basis for vocal registers, revealing that a well-developed ligament is not needed for loft-modal quality shifts as seen in harmonic amplitude measures. Results also reveal that a distinctively pulsatile register can occur in infants at a much higher f o than expected on psychoacoustic grounds. Overall results are consistent with cover tissues in infancy that are, for vibratory purposes, highly compliant and readily detached. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Cook, Perry
This chapter covers algorithms, technologies, computer languages, and systems for computer music. Computer music involves the application of computers and other digital/electronic technologies to music composition, performance, theory, history, and perception. The field combines digital signal processing, computational algorithms, computer languages, hardware and software systems, acoustics, psychoacoustics (low-level perception of sounds from the raw acoustic signal), and music cognition (higher-level perception of musical style, form, emotion, etc.). Although most people would think that analog synthesizers and electronic music substantially predate the use of computers in music, many experiments and complete computer music systems were being constructed and used as early as the 1950s.
Calculation of selective filters of a device for primary analysis of speech signals
NASA Astrophysics Data System (ADS)
Chudnovskii, L. S.; Ageev, V. M.
2014-07-01
The amplitude-frequency responses of filters for primary analysis of speech signals, which have a low quality factor and a high rolloff factor in the high-frequency range, are calculated using the linear theory of speech production and psychoacoustic measurement data. The frequency resolution of the filter system for a sinusoidal signal is 40-200 Hz. The modulation-frequency resolution of amplitude- and frequency-modulated signals is 3-6 Hz. The aforementioned features of the calculated filters are close to the amplitudefrequency responses of biological auditory systems at the level of the eighth nerve.
USAF Bioenvironmental Noise Data Handbook. Volume 167: MA-3M air conditioner
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-06-01
The MA-3M is an electric motor-driven air conditioner designed to cool electronic equipment on aircraft during ground maintenance. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating at a normal rated condition. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference levels, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors.
USAF Bioenvironmental Noise Data Handbook. Volume 149: C-9A in-flight crew/passenger noise
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The C-9A is a McDonnell Douglas DC-9 series 30 commercial transport modified to perform aeromedical evacuation missions. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this aircraft during normal flight operations. Data are reported for 56 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors.
Auditory models for speech analysis
NASA Astrophysics Data System (ADS)
Maybury, Mark T.
This paper reviews the psychophysical basis for auditory models and discusses their application to automatic speech recognition. First an overview of the human auditory system is presented, followed by a review of current knowledge gleaned from neurological and psychoacoustic experimentation. Next, a general framework describes established peripheral auditory models which are based on well-understood properties of the peripheral auditory system. This is followed by a discussion of current enhancements to that models to include nonlinearities and synchrony information as well as other higher auditory functions. Finally, the initial performance of auditory models in the task of speech recognition is examined and additional applications are mentioned.
Factors Affecting Acoustics and Speech Intelligibility in the Operating Room: Size Matters
Bennett, Christopher L.; Horn, Danielle Bodzin; Dudaryk, Roman
2017-01-01
INTRODUCTION: Noise in health care settings has increased since 1960 and represents a significant source of dissatisfaction among staff and patients and risk to patient safety. Operating rooms (ORs) in which effective communication is crucial are particularly noisy. Speech intelligibility is impacted by noise, room architecture, and acoustics. For example, sound reverberation time (RT60) increases with room size, which can negatively impact intelligibility, while room objects are hypothesized to have the opposite effect. We explored these relationships by investigating room construction and acoustics of the surgical suites at our institution. METHODS: We studied our ORs during times of nonuse. Room dimensions were measured to calculate room volumes (VR). Room content was assessed by estimating size and assigning items into 5 volume categories to arrive at an adjusted room content volume (VC) metric. Psychoacoustic analyses were performed by playing sweep tones from a speaker and recording the impulse responses (ie, resulting sound fields) from 3 locations in each room. The recordings were used to calculate 6 psychoacoustic indices of intelligibility. Multiple linear regression was performed using VR and VC as predictor variables and each intelligibility index as an outcome variable. RESULTS: A total of 40 ORs were studied. The surgical suites were characterized by a large degree of construction and surface finish heterogeneity and varied in size from 71.2 to 196.4 m3 (average VR = 131.1 [34.2] m3). An insignificant correlation was observed between VR and VC (Pearson correlation = 0.223, P = .166). Multiple linear regression model fits and β coefficients for VR were highly significant for each of the intelligibility indices and were best for RT60 (R2 = 0.666, F(2, 37) = 39.9, P < .0001). For Dmax (maximum distance where there is <15% loss of consonant articulation), both VR and VC β coefficients were significant. For RT60 and Dmax, after controlling for VC, partial correlations were 0.825 (P < .0001) and 0.718 (P < .0001), respectively, while after controlling for VR, partial correlations were −0.322 (P = .169) and 0.381 (P < .05), respectively. CONCLUSIONS: Our results suggest that the size and contents of an OR can predict a range of psychoacoustic indices of speech intelligibility. Specifically, increasing OR size correlated with worse speech intelligibility, while increasing amounts of OR contents correlated with improved speech intelligibility. This study provides valuable descriptive data and a predictive method for identifying existing ORs that may benefit from acoustic modifiers (eg, sound absorption panels). Additionally, it suggests that room dimensions and projected clinical use should be considered during the design phase of OR suites to optimize acoustic performance. PMID:28525511
Psychoacoustic Factors in Musical Intonation: Beats, Interval Tuning, and Inharmonicity.
NASA Astrophysics Data System (ADS)
Keislar, Douglas Fleming
Three psychoacoustic experiments were conducted using musically experienced subjects. In the first two experiments, the interval tested was the perfect fifth F4-C5; in the final one it was the major third F4-A4. The beat rate was controlled by two different methods: (1) simply retuning the interval, and (2) frequency-shifting one partial of each pair of beating partials without changing the overall interval tuning. The second method introduces inharmonicity. In addition, two levels of beat amplitude were introduced by using either a complete spectrum of 16 equal-amplitude partials per note, or by deleting one partial from each pair of beating partials. The results of all three experiments indicate that, for these stimuli, beating does not contribute significantly to the percept of "out-of-tuneness," because it made no difference statistically whether the beat amplitude was maximal or minimal. By contrast, mistuning the interval was highly significant. For the fifths, frequency-shifting the appropriate partials had about as much effect on the perceived intonation as mistuning the interval. For thirds, this effect was weaker, presumably since there were fewer inharmonic partials and they were higher in the harmonic series. Subjects were less consistent in their judgments of thirds than of fifths, perhaps because the equal-tempered and just thirds differ noticeably, unlike fifths. Since it is unlikely that beats would be more audible in real musical situations than under these laboratory conditions, these results suggest that the perception of intonation in music is dependent on the actual interval tuning rather than the concomitant beat rate. If beating partials are unimportant vis-a-vis interval tuning, this strengthens the argument for a cultural basis for musical intonation and scales, as opposed to the acoustical basis set forth by Helmholtz and others.
O'Sullivan, James A; Shamma, Shihab A; Lalor, Edmund C
2015-05-06
The human brain has evolved to operate effectively in highly complex acoustic environments, segregating multiple sound sources into perceptually distinct auditory objects. A recent theory seeks to explain this ability by arguing that stream segregation occurs primarily due to the temporal coherence of the neural populations that encode the various features of an individual acoustic source. This theory has received support from both psychoacoustic and functional magnetic resonance imaging (fMRI) studies that use stimuli which model complex acoustic environments. Termed stochastic figure-ground (SFG) stimuli, they are composed of a "figure" and background that overlap in spectrotemporal space, such that the only way to segregate the figure is by computing the coherence of its frequency components over time. Here, we extend these psychoacoustic and fMRI findings by using the greater temporal resolution of electroencephalography to investigate the neural computation of temporal coherence. We present subjects with modified SFG stimuli wherein the temporal coherence of the figure is modulated stochastically over time, which allows us to use linear regression methods to extract a signature of the neural processing of this temporal coherence. We do this under both active and passive listening conditions. Our findings show an early effect of coherence during passive listening, lasting from ∼115 to 185 ms post-stimulus. When subjects are actively listening to the stimuli, these responses are larger and last longer, up to ∼265 ms. These findings provide evidence for early and preattentive neural computations of temporal coherence that are enhanced by active analysis of an auditory scene. Copyright © 2015 the authors 0270-6474/15/357256-08$15.00/0.
Digital Sound Synthesis Algorithms: a Tutorial Introduction and Comparison of Methods
NASA Astrophysics Data System (ADS)
Lee, J. Robert
The objectives of the dissertation are to provide both a compendium of sound-synthesis methods with detailed descriptions and sound examples, as well as a comparison of the relative merits of each method based on ease of use, observed sound quality, execution time, and data storage requirements. The methods are classified under the general headings of wavetable-lookup synthesis, additive synthesis, subtractive synthesis, nonlinear methods, and physical modelling. The nonlinear methods comprise a large group that ranges from the well-known frequency-modulation synthesis to waveshaping. The final category explores computer modelling of real musical instruments and includes numerical and analytical solutions to the classical wave equation of motion, along with some of the more sophisticated time -domain models that are possible through the prudent combination of simpler synthesis techniques. The dissertation is intended to be understandable by a musician who is mathematically literate but who does not necessarily have a background in digital signal processing. With this limitation in mind, a brief and somewhat intuitive description of digital sampling theory is provided in the introduction. Other topics such as filter theory are discussed as the need arises. By employing each of the synthesis methods to produce the same type of sound, interesting comparisons can be made. For example, a struck string sound, such as that typical of a piano, can be produced by algorithms in each of the synthesis classifications. Many sounds, however, are peculiar to a single algorithm and must be examined independently. Psychoacoustic studies were conducted as an aid in the comparison of the sound quality of several implementations of the synthesis algorithms. Other psychoacoustic experiments were conducted to supplement the established notions of which timbral issues are important in the re -synthesis of the sounds of acoustic musical instruments.
Neuromechanistic Model of Auditory Bistability
Rankin, James; Sussman, Elyse; Rinzel, John
2015-01-01
Sequences of higher frequency A and lower frequency B tones repeating in an ABA- triplet pattern are widely used to study auditory streaming. One may experience either an integrated percept, a single ABA-ABA- stream, or a segregated percept, separate but simultaneous streams A-A-A-A- and -B---B--. During minutes-long presentations, subjects may report irregular alternations between these interpretations. We combine neuromechanistic modeling and psychoacoustic experiments to study these persistent alternations and to characterize the effects of manipulating stimulus parameters. Unlike many phenomenological models with abstract, percept-specific competition and fixed inputs, our network model comprises neuronal units with sensory feature dependent inputs that mimic the pulsatile-like A1 responses to tones in the ABA- triplets. It embodies a neuronal computation for percept competition thought to occur beyond primary auditory cortex (A1). Mutual inhibition, adaptation and noise are implemented. We include slow NDMA recurrent excitation for local temporal memory that enables linkage across sound gaps from one triplet to the next. Percepts in our model are identified in the firing patterns of the neuronal units. We predict with the model that manipulations of the frequency difference between tones A and B should affect the dominance durations of the stronger percept, the one dominant a larger fraction of time, more than those of the weaker percept—a property that has been previously established and generalized across several visual bistable paradigms. We confirm the qualitative prediction with our psychoacoustic experiments and use the behavioral data to further constrain and improve the model, achieving quantitative agreement between experimental and modeling results. Our work and model provide a platform that can be extended to consider other stimulus conditions, including the effects of context and volition. PMID:26562507
ERP evidence for the recognition of emotional prosody through simulated cochlear implant strategies.
Agrawal, Deepashri; Timm, Lydia; Viola, Filipa Campos; Debener, Stefan; Büchner, Andreas; Dengler, Reinhard; Wittfoth, Matthias
2012-09-20
Emotionally salient information in spoken language can be provided by variations in speech melody (prosody) or by emotional semantics. Emotional prosody is essential to convey feelings through speech. In sensori-neural hearing loss, impaired speech perception can be improved by cochlear implants (CIs). Aim of this study was to investigate the performance of normal-hearing (NH) participants on the perception of emotional prosody with vocoded stimuli. Semantically neutral sentences with emotional (happy, angry and neutral) prosody were used. Sentences were manipulated to simulate two CI speech-coding strategies: the Advance Combination Encoder (ACE) and the newly developed Psychoacoustic Advanced Combination Encoder (PACE). Twenty NH adults were asked to recognize emotional prosody from ACE and PACE simulations. Performance was assessed using behavioral tests and event-related potentials (ERPs). Behavioral data revealed superior performance with original stimuli compared to the simulations. For simulations, better recognition for happy and angry prosody was observed compared to the neutral. Irrespective of simulated or unsimulated stimulus type, a significantly larger P200 event-related potential was observed for happy prosody after sentence onset than the other two emotions. Further, the amplitude of P200 was significantly more positive for PACE strategy use compared to the ACE strategy. Results suggested P200 peak as an indicator of active differentiation and recognition of emotional prosody. Larger P200 peak amplitude for happy prosody indicated importance of fundamental frequency (F0) cues in prosody processing. Advantage of PACE over ACE highlighted a privileged role of the psychoacoustic masking model in improving prosody perception. Taken together, the study emphasizes on the importance of vocoded simulation to better understand the prosodic cues which CI users may be utilizing.
Pitch sensation involves stochastic resonance
Martignoli, Stefan; Gomez, Florian; Stoop, Ruedi
2013-01-01
Pitch is a complex hearing phenomenon that results from elicited and self-generated cochlear vibrations. Read-off vibrational information is relayed higher up the auditory pathway, where it is then condensed into pitch sensation. How this can adequately be described in terms of physics has largely remained an open question. We have developed a peripheral hearing system (in hardware and software) that reproduces with great accuracy all salient pitch features known from biophysical and psychoacoustic experiments. At the level of the auditory nerve, the system exploits stochastic resonance to achieve this performance, which may explain the large amount of noise observed in the working auditory nerve. PMID:24045830
Design, construction, and evaluation of a 1:8 scale model binaural manikin.
Robinson, Philip; Xiang, Ning
2013-03-01
Many experiments in architectural acoustics require presenting listeners with simulations of different rooms to compare. Acoustic scale modeling is a feasible means to create accurate simulations of many rooms at reasonable cost. A critical component in a scale model room simulation is a receiver that properly emulates a human receiver. For this purpose, a scale model artificial head has been constructed and tested. This paper presents the design and construction methods used, proper equalization procedures, and measurements of its response. A headphone listening experiment examining sound externalization with various reflection conditions is presented that demonstrates its use for psycho-acoustic testing.
Challenges Facing 3-D Audio Display Design for Multimedia
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Null, Cynthia H. (Technical Monitor)
1998-01-01
The challenges facing successful multimedia presentation depend largely on the expectations of the designer and end user for a given application. Perceptual limitations in distance, elevation and azimuth sound source simulation differ significantly between headphone and cross-talk cancellation loudspeaker listening and therefore must be considered. Simulation of an environmental context is desirable but the quality depends on processing resources and lack of interaction with the host acoustical environment. While techniques such as data reduction of head-related transfer functions have been used widely to improve simulation fidelity, another approach involves determining thresholds for environmental acoustic events. Psychoacoustic studies relevant to this approach are reviewed in consideration of multimedia applications
UDAF bioenvironmental noise data handbook. Volume 164: MD-1 heater, duct type, portable
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The MD-1 heater is an electric motor-driven, portable ground heater used primarily for cockpit and cabin temperature control. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at normal rated conditions. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors.
USAF bioenvironmental noise data handbook. Volume 162: MD-4MO generator set
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The MD-4MO generator set is an electric motor-driven source of electrical power used primarily for the starting of aircraft, and for ground maintenance. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at a normal rated condition. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference levels, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors.
Whistle register: a preliminary investigation by HSDI visualization and acoustics on female cases
NASA Astrophysics Data System (ADS)
Di Corcia, Antonio; Fussi, Franco
2012-02-01
In this study we investigated laryngeal behaviors involved during vocal production of highest female vocal ranges in Flute in M3 Register, in Whistle Register and in a newly formulated by us, Hiss Register. Observations were carried with stroboscopy and High Speed Digital Imaging and with spectrographic and psycho-acoustic analysis by means of a software system having a wide spectral range (0-20.000 Hz). Results indicate that at the highest pitch vocal folds vibration is absent or significantly reduced, glottic contact is incomplete. These acoustic form of extreme pitch levels comprised intra-harmonic noise and overtones within 10 to 18 kHz range.
Binaural fusion and the representation of virtual pitch in the human auditory cortex.
Pantev, C; Elbert, T; Ross, B; Eulitz, C; Terhardt, E
1996-10-01
The auditory system derives the pitch of complex tones from the tone's harmonics. Research in psychoacoustics predicted that binaural fusion was an important feature of pitch processing. Based on neuromagnetic human data, the first neurophysiological confirmation of binaural fusion in hearing is presented. The centre of activation within the cortical tonotopic map corresponds to the location of the perceived pitch and not to the locations that are activated when the single frequency constituents are presented. This is also true when the different harmonics of a complex tone are presented dichotically. We conclude that the pitch processor includes binaural fusion to determine the particular pitch location which is activated in the auditory cortex.
A centralized audio presentation manager
DOE Office of Scientific and Technical Information (OSTI.GOV)
Papp, A.L. III; Blattner, M.M.
1994-05-16
The centralized audio presentation manager addresses the problems which occur when multiple programs running simultaneously attempt to use the audio output of a computer system. Time dependence of sound means that certain auditory messages must be scheduled simultaneously, which can lead to perceptual problems due to psychoacoustic phenomena. Furthermore, the combination of speech and nonspeech audio is examined; each presents its own problems of perceptibility in an acoustic environment composed of multiple auditory streams. The centralized audio presentation manager receives abstract parameterized message requests from the currently running programs, and attempts to create and present a sonic representation in themore » most perceptible manner through the use of a theoretically and empirically designed rule set.« less
A new version of the helicopter aural detection program, ICHIN
NASA Technical Reports Server (NTRS)
Mueller, A. W.; Smith, C. D.; Shepherd, K. P.; Sullivan, B. M.
1986-01-01
NASA Langley Research Center personnel have conducted an evaluation of the helicopter aural detection program I Can Hear It Now (ICHIN version-5). This was accomplished using flight noise data of five helicopters, obtained from a joint NASA and U.S. Army acoustics measurement program. The evaluation consisted of presenting the noise data to a jury of 20 subjects and to the ICHIN-5 program. A comparative study was then made of the detection distances determined by the jury and predicted by ICHIN-5. This report presents the changes made in the ICHIN-5 program as a result of this comparative study. The changes represent current psychoacoustics and propagation knowledge.
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-06-01
The MB-3 Tester is an electric motor-driven cabin leakage tester designed to furnish pressurized air to the aircraft at controlled pressures and temperatures during ground pressurization of aircraft cockpits and pressurized compartments. This report provides measured data defining the bioacoustic environments produced by this unit operating at a normal rated/load condition. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors.
Rodent Auditory Perception: Critical Band Limitations and Plasticity
King, Julia; Insanally, Michele; Jin, Menghan; Martins, Ana Raquel O.; D'amour, James A.; Froemke, Robert C.
2015-01-01
What do animals hear? While it remains challenging to adequately assess sensory perception in animal models, it is important to determine perceptual abilities in model systems to understand how physiological processes and plasticity relate to perception, learning, and cognition. Here we discuss hearing in rodents, reviewing previous and recent behavioral experiments querying acoustic perception in rats and mice, and examining the relation between behavioral data and electrophysiological recordings from the central auditory system. We focus on measurements of critical bands, which are psychoacoustic phenomena that seem to have a neural basis in the functional organization of the cochlea and the inferior colliculus. We then discuss how behavioral training, brain stimulation, and neuropathology impact auditory processing and perception. PMID:25827498
Annoyance to Noise Produced by a Distributed Electric Propulsion High-Lift System
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.; Palumbo, Daniel L.; Rathsam, Jonathan; Christian, Andrew; Rafaelof, Menachem
2017-01-01
A psychoacoustic test was performed using simulated sounds from a distributed electric propulsion aircraft concept to help understand factors associated with human annoyance. A design space spanning the number of high-lift leading edge propellers and their relative operating speeds, inclusive of time varying effects associated with motor controller error and atmospheric turbulence, was considered. It was found that the mean annoyance response varies in a statistically significant manner with the number of propellers and with the inclusion of time varying effects, but does not differ significantly with the relative RPM between propellers. An annoyance model was developed, inclusive of confidence intervals, using the noise metrics of loudness, roughness, and tonality as predictors.
Govaerts, Paul J; Vaerenberg, Bart; De Ceulaer, Geert; Daemers, Kristin; De Beukelaer, Carina; Schauwers, Karen
2010-08-01
An intelligent agent, Fitting to Outcomes eXpert, was developed to optimize and automate Cochlear implant (CI) programming. The current article describes the rationale, development, and features of this tool. Cochlear implant fitting is a time-consuming procedure to define the value of a subset of the available electric parameters based primarily on behavioral responses. It is comfort-driven with high intraindividual and interindividual variability both with respect to the patient and to the clinician. Its validity in terms of process control can be questioned. Good clinical practice would require an outcome-driven approach. An intelligent agent may help solve the complexity of addressing more electric parameters based on a range of outcome measures. A software application was developed that consists of deterministic rules that analyze the map settings in the processor together with psychoacoustic test results (audiogram, A(section sign)E phoneme discrimination, A(section sign)E loudness scaling, speech audiogram) obtained with that map. The rules were based on the daily clinical practice and the expertise of the CI programmers. The data transfer to and from this agent is either manual or through seamless digital communication with the CI fitting database and the psychoacoustic test suite. It recommends and executes modifications to the map settings to improve the outcome. Fitting to Outcomes eXpert is an operational intelligent agent, the principles of which are described. Its development and modes of operation are outlined, and a case example is given. Fitting to Outcomes eXpert is in use for more than a year now and seems to be capable to improve the measured outcome. It is argued that this novel tool allows a systematic approach focusing on outcome, reducing the fitting time, and improving the quality of fitting. It introduces principles of artificial intelligence in the process of CI fitting.
Assessment of noise metrics for application to rotorcraft
NASA Astrophysics Data System (ADS)
McMullen, Andrew L.
It is anticipated that the use of rotorcraft passenger vehicles for shorter journeys will increase because their use can reduce the time between boarding and take-off. The characteristics of rotorcraft noise are very different to that of fixed wing aircraft. There can be strong tonal components, fluctuations that can also make the noise sound impulsive, and future rotorcraft may produce proportionally more low frequency noise content. Most metrics that are used today to predict noise impact on communities around airports (e.g., Ldn) are just functions of A-weighted sound pressure level. To build a better noise annoyance model that can be applied to assess impact of future and current rotorcraft, it is important to understand the perceived sound attributes and how they influence annoyance. A series of psychoacoustic tests were designed and performed to further our understanding of how rotorcraft sound characteristics affect annoyance as well as evaluate the applicability of existing noise metrics as predictors of annoyance due to rotorcraft noise. The effect of the method used to reproduce sounds in the psychoacoustics tests was also investigated, and so tests were conducted in the NASA Langley Exterior Effects Room using loudspeaker arrays to simulate flyovers and in a double walled sound booth using earphones for playback. A semantic differential test was performed, and analysis of subject responses showed the presence of several independent perceptual factors relating to: loudness, sharpness, roughness, tonality, and impulsiveness. A simulation method was developed to alter tonal components in existing rotorcraft flyover recordings to change the impulsiveness and tonality of the sounds. Flyover recordings and simulations with varied attributes were used as stimuli in an annoyance test. Results showed that EPNL and SELA performed well as predictors of annoyance, but outliers to generate trends have tonal related characteristics that could be contributing to annoyance. General trends in results were similar for both test environments, though differences were greater for the annoyance tests than the semantic differential tests.
Williamson, Victoria J; Stewart, Lauren
2013-01-01
For most people, music, like language, is acquired effortlessly in early life. But a few percent of the population have lifelong difficulties in the perception and production of music. In this chapter we discuss psycho-acoustic and behavioral studies that have attempted to delineate the nature of the auditory perceptual deficits in this group and consider whether these difficulties extend outside the musical domain. Finally, we review structural imaging studies in this group which point to subtle anomalies in temporal and frontal areas. We suggest that amusia can be considered a disorder of neural development, which has relatively specific consequences at the behavioral level. Studies of congenital amusia provide a unique window on the neurocognitive architecture of music processing. Copyright © 2013 Elsevier B.V. All rights reserved.
Emergence of collective intonation in the musical performance of crowds
NASA Astrophysics Data System (ADS)
Lacasa, Lucas
2016-09-01
The average individual is typically a mediocre singer, with a rather restricted capacity to sing a melody in tune. Yet when many singers are assembled to perform collectively, the resulting melody of the crowd is suddenly perceived by an external listener as perfectly tuned —as if it was actually a choral performance— even if each individual singer is out of tune. This collective phenomenon is an example of a wisdom of crowds effect that can be routinely observed in music concerts or other social events, when a group of people spontaneously sings at unison. In this paper we rely on the psychoacoustic properties of pitch and provide a simple mechanistic explanation for the onset of this emergent behavior.
[The physicist Félix Savart (1791-1841). Physician/Surgeon, pioneer in the study of acoustics].
Ségal, Alain
2015-01-01
Félix Savart (1791-1841) was both a physician and a physicist, and also a pioneer of acoustics and psycho-acoustics. In 1819 Savart scientifically devised and contructed a trapezoidal violin with the advice of Paris string-instrument maker J-B Vuillaume. This violin drew the attention of J-B Biot who suggested young Savart to work with him on such acoustic researches. From this collaboration proceeded the so-called "law of Biot-Savart" about magnetic power which was in fact formulated by Pierre Simon de Laplace (1749-1827). Savart worked on numerous and diversified acoustic researches. Lord John Rayleigh described them as "beautiful experiments", and he became a foreign correspondant member of the Royal Society in 1839.
A Study in a New Test Facility on Indoor Annoyance Caused by Sonic Booms
NASA Technical Reports Server (NTRS)
Rathsam, Jonathan; Loubeau, Alexandra; Klos, Jacob
2012-01-01
A sonic-boom simulator at NASA Langley Research Center has been constructed to research the indoor human response to low-amplitude sonic booms. The research goal is the development of a psychoacoustic model for individual sonic booms to be validated by future community studies. The study in this report assessed the suitability of existing noise metrics for predicting indoor human annoyance. The test signals included a wide range of synthesized and recorded sonic-boom waveforms. Results indicated that no noise metric predicts indoor annoyance to sonic-boom sounds better than Perceived Level, PL. During the study it became apparent that structural vibrations induced by the test signals were contributing to annoyance, so the relationship between sound and vibration at levels of equivalent annoyance has been quantified.
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Null, Cynthia H. (Technical Monitor)
1997-01-01
This talk will overview the basic technologies related to the creation of virtual acoustic images, and the potential of including spatial auditory displays in human-machine interfaces. Research into the perceptual error inherent in both natural and virtual spatial hearing is reviewed, since the formation of improved technologies is tied to psychoacoustic research. This includes a discussion of Head Related Transfer Function (HRTF) measurement techniques (the HRTF provides important perceptual cues within a virtual acoustic display). Many commercial applications of virtual acoustics have so far focused on games and entertainment ; in this review, other types of applications are examined, including aeronautic safety, voice communications, virtual reality, and room acoustic simulation. In particular, the notion that realistic simulation is optimized within a virtual acoustic display when head motion and reverberation cues are included within a perceptual model.
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-07-01
Measured and extrapolated data define the bioacoustic environments produced by a gasoline engine driven cabin leakage tester operating outdoors on a concrete apron at normal rated conditions. Near field data are presented for 37 locations at a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 - 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
Variations on a theme by chopin: relations between perception and production of timing in music.
Repp, B H
1998-06-01
A note interonset interval (IOI) increment in mechanically timed music is more difficult to detect where expressive lengthening typically occurs in artistic performance. Experiment 1 showed this in an excerpt from a Chopin etude and extended the task to IOI decrement detection. A simple measure of variation in perceptual bias was derived that correlated highly with the average timing pattern of pianists' performances, more so than with acoustic surface properties of the music. Similar results, but decreasing correlations, were obtained in each of four subsequent experiments in which the music was simplified in stages. Although local psychoacoustic effects on time perception cannot be ruled out completely, the results suggest that musical structure (melodic-rhythmic grouping in particular) has temporal implications that are reflected not only in musicians' motor behavior but also in listeners' time-keeping abilities.
Generating partially correlated noise—A comparison of methods
Hartmann, William M.; Cho, Yun Jin
2011-01-01
There are three standard methods for generating two channels of partially correlated noise: the two-generator method, the three-generator method, and the symmetric-generator method. These methods allow an experimenter to specify a target cross correlation between the two channels, but actual generated noises show statistical variability around the target value. Numerical experiments were done to compare the variability for those methods as a function of the number of degrees of freedom. The results of the experiments quantify the stimulus uncertainty in diverse binaural psychoacoustical experiments: incoherence detection, perceived auditory source width, envelopment, noise localization∕lateralization, and the masking level difference. The numerical experiments found that when the elemental generators have unequal powers, the different methods all have similar variability. When the powers are constrained to be equal, the symmetric-generator method has much smaller variability than the other two. PMID:21786899
Interaural time sensitivity of high-frequency neurons in the inferior colliculus.
Yin, T C; Kuwada, S; Sujaku, Y
1984-11-01
Recent psychoacoustic experiments have shown that interaural time differences provide adequate cues for lateralizing high-frequency sounds, provided the stimuli are complex and not pure tones. We present here physiological evidence in support of these findings. Neurons of high best frequency in the cat inferior colliculus respond to interaural phase differences of amplitude modulated waveforms, and this response depends upon preservation of phase information of the modulating signal. Interaural phase differences were introduced in two ways: by interaural delays of the entire waveform and by binaural beats in which there was an interaural frequency difference in the modulating waveform. Results obtained with these two methods are similar. Our results show that high-frequency cells can respond to interaural time differences of amplitude modulated signals and that they do so by a sensitivity to interaural phase differences of the modulating waveform.
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Godfroy, Martine; Sandor, Aniko; Holden, Kritina
2008-01-01
The design of caution-warning signals for NASA s Crew Exploration Vehicle (CEV) and other future spacecraft will be based on both best practices based on current research and evaluation of current alarms. A design approach is presented based upon cross-disciplinary examination of psychoacoustic research, human factors experience, aerospace practices, and acoustical engineering requirements. A listening test with thirteen participants was performed involving ranking and grading of current and newly developed caution-warning stimuli under three conditions: (1) alarm levels adjusted for compliance with ISO 7731, "Danger signals for work places - Auditory Danger Signals", (2) alarm levels adjusted to an overall 15 dBA s/n ratio and (3) simulated codec low-pass filtering. Questionnaire data yielded useful insights regarding cognitive associations with the sounds.
USAF Bioenvironmental Noise Data Handbook. Volume 155. CH-3 in-flight crew noise
NASA Astrophysics Data System (ADS)
Hille, H. K.
1982-09-01
The CH-3 is a USAF tactical combat transport helicopter. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this helicopter during normal flight operations. Data are reported for nine locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C weighted and A weighted sound levels, preferred speech interference level, perceived noise levels and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Refer to Volume 1 of this handbook, USAF Bioenvironmental Noise Data handbook, Vol. 1: Organization, Content and Application, AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
USAF Environmental Noise Data Handbook. Volume 150: C-140 in-flight crew noise
NASA Astrophysics Data System (ADS)
Hille, H. K.
1982-09-01
The C-140 is a USAF transport aircraft used for operational support. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this aircraft during normal flight operations. Date are reported for seven locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Refer to Volume 1 of this handbook, USAF Bioenvironmental Noise Data Handbook, Vol. 1: Organization, Content and Application, AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
The elements of a comprehensive education for future architectural acousticians
NASA Astrophysics Data System (ADS)
Wang, Lily M.
2005-04-01
Curricula for students who seek to become consultants of architectural acoustics or researchers in the field are few in the United States and in the world. This paper will present the author's opinions on the principal skills a student should obtain from a focused course of study in architectural acoustics. These include: (a) a solid command of math and wave theory, (b) fluency with digital signal processing techniques and sound measurement equipment, (c) expertise in using architectural acoustic software with an understanding of its limitations, (d) knowledge of building mechanical systems, (e) an understanding of human psychoacoustics, and (f) an appreciation for the artistic aspects of the discipline. Additionally, writing and presentation skills should be emphasized and participation in professional societies encouraged. Armed with such abilities, future architectural acousticians will advance the field significantly.
State of the art in perceptual design of hearing aids
NASA Astrophysics Data System (ADS)
Edwards, Brent W.; van Tasell, Dianne J.
2002-05-01
Hearing aid capabilities have increased dramatically over the past six years, in large part due to the development of small, low-power digital signal processing chips suitable for hearing aid applications. As hearing aid signal processing capabilities increase, there will be new opportunities to apply perceptually based knowledge to technological development. Most hearing loss compensation techniques in today's hearing aids are based on simple estimates of audibility and loudness. As our understanding of the psychoacoustical and physiological characteristics of sensorineural hearing loss improves, the result should be improved design of hearing aids and fitting methods. The state of the art in hearing aids will be reviewed, including form factors, user requirements, and technology that improves speech intelligibility, sound quality, and functionality. General areas of auditory perception that remain unaddressed by current hearing aid technology will be discussed.
Allgood, Rebecca; Heaton, Pamela
2015-09-01
Although the configurations of psychoacoustic cues signalling emotions in human vocalizations and instrumental music are very similar, cross-domain links in recognition performance have yet to be studied developmentally. Two hundred and twenty 5- to 10-year-old children were asked to identify musical excerpts and vocalizations as happy, sad, or fearful. The results revealed age-related increases in overall recognition performance with significant correlations across vocal and musical conditions at all developmental stages. Recognition scores were greater for musical than vocal stimuli and were superior in females compared with males. These results confirm that recognition of emotions in vocal and musical stimuli is linked by 5 years and that sensitivity to emotions in auditory stimuli is influenced by age and gender. © 2015 The British Psychological Society.
Categorization of extremely brief auditory stimuli: domain-specific or domain-general processes?
Bigand, Emmanuel; Delbé, Charles; Gérard, Yannick; Tillmann, Barbara
2011-01-01
The present study investigated the minimum amount of auditory stimulation that allows differentiation of spoken voices, instrumental music, and environmental sounds. Three new findings were reported. 1) All stimuli were categorized above chance level with 50 ms-segments. 2) When a peak-level normalization was applied, music and voices started to be accurately categorized with 20 ms-segments. When the root-mean-square (RMS) energy of the stimuli was equalized, voice stimuli were better recognized than music and environmental sounds. 3) Further psychoacoustical analyses suggest that the categorization of extremely brief auditory stimuli depends on the variability of their spectral envelope in the used set. These last two findings challenge the interpretation of the voice superiority effect reported in previously published studies and propose a more parsimonious interpretation in terms of an emerging property of auditory categorization processes.
A keyword spotting model using perceptually significant energy features
NASA Astrophysics Data System (ADS)
Umakanthan, Padmalochini
The task of a keyword recognition system is to detect the presence of certain words in a conversation based on the linguistic information present in human speech. Such keyword spotting systems have applications in homeland security, telephone surveillance and human-computer interfacing. General procedure of a keyword spotting system involves feature generation and matching. In this work, new set of features that are based on the psycho-acoustic masking nature of human speech are proposed. After developing these features a time aligned pattern matching process was implemented to locate the words in a set of unknown words. A word boundary detection technique based on frame classification using the nonlinear characteristics of speech is also addressed in this work. Validation of this keyword spotting model was done using widely acclaimed Cepstral features. The experimental results indicate the viability of using these perceptually significant features as an augmented feature set in keyword spotting.
Fractal dimension analysis of complexity in Ligeti piano pieces
NASA Astrophysics Data System (ADS)
Bader, Rolf
2005-04-01
Fractal correlation dimensional analysis has been performed with whole solo piano pieces by Gyrgy Ligeti at every 50ms interval of the pieces. The resulting curves of development of complexity represented by the fractal dimension showed up a very reasonable correlation with the perceptional density of events during these pieces. The seventh piece of Ligeti's ``Musica ricercata'' was used as a test case. Here, each new part of the piece was followed by an increase of the fractal dimension because of the increase of information at the part changes. The second piece ``Galamb borong,'' number seven of the piano Etudes was used, because Ligeti wrote these Etudes after studying fractal geometry. Although the piece is not fractal in the strict mathematical sense, the overall structure of the psychoacoustic event-density as well as the detailed event development is represented by the fractal dimension plot.
Generating partially correlated noise--a comparison of methods.
Hartmann, William M; Cho, Yun Jin
2011-07-01
There are three standard methods for generating two channels of partially correlated noise: the two-generator method, the three-generator method, and the symmetric-generator method. These methods allow an experimenter to specify a target cross correlation between the two channels, but actual generated noises show statistical variability around the target value. Numerical experiments were done to compare the variability for those methods as a function of the number of degrees of freedom. The results of the experiments quantify the stimulus uncertainty in diverse binaural psychoacoustical experiments: incoherence detection, perceived auditory source width, envelopment, noise localization/lateralization, and the masking level difference. The numerical experiments found that when the elemental generators have unequal powers, the different methods all have similar variability. When the powers are constrained to be equal, the symmetric-generator method has much smaller variability than the other two. © 2011 Acoustical Society of America
Validation of a Korean Version of the Tinnitus Handicap Questionnaire
Yoo, Ik Won; Hwang, Sun Jin; Hwang, Soon Young
2015-01-01
Objectives The goal of the present study was to evaluate the reliability and validity of the Korean version of the tinnitus handicap questionnaire (THQ-K). Methods A total of 60 patients were included in this study. Patients responded to the THQ-K, the tinnitus handicap inventory (THI), Beck's depression index (BDI), and the visual analogue scale (VAS) for loudness and pitch, loudness match, and minimum masking level (MML) test were performed. Results Internal consistency of the THQ-K was examined using Cronbach coefficient alpha. Cronbach alpha was 0.96. The THQ-K showed a significant correlation with THI, BDI, VAS for distress, and VAS for loudness, but no significant correlation with psychoacoustic measurement of tinnitus, such as loudness match, pitch match, and MML. Conclusion The THQ-K is a reliable and valid test for evaluating the degree of handicap due to tinnitus for both research and clinical use. PMID:26330911
NASA Astrophysics Data System (ADS)
Hubbard, H. H.; Powell, C. A.
1981-06-01
A number of facilities were developed which provide a unique test capability for psychoacoustics and related human factors research. The design philosophy, physical layouts, dimensions, construction features, operating capabilities, and example applications for these facilities are described. In the exterior effects room, human subjects are exposed to the types of noises that are experienced outdoors, and in the interior effects room, subjects are exposed to the types of noises and noise-induced vibrations that are experience indoors. Subjects are also exposed to noises in an echo-free environment in the anechoic listening room. An aircraft noise synthesis system, which simulates aircraft flyover noise at an observer position on the ground, is used in conjunction with these three rooms. The passenger ride quality apparatus, a device for studying passenger response to noise and vibration in aircraft, or in other vehicles, is described.
Development and evaluation of a general aviation real world noise simulator
NASA Technical Reports Server (NTRS)
Galanter, E.; Popper, R.
1980-01-01
An acoustic playback system is described which realistically simulates the sounds experienced by the pilot of a general aviation aircraft during engine idle, take-off, climb, cruise, descent, and landing. The physical parameters of the signal as they appear in the simulator environment are compared to analogous parameters derived from signals recorded during actual flight operations. The acoustic parameters of the simulated and real signals during cruise conditions are within plus or minus two dB in third octave bands from 0.04 to 4 kHz. The overall A-weighted levels of the signals are within one dB of signals generated in the actual aircraft during equivalent maneuvers. Psychoacoustic evaluations of the simulator signal are compared with similar measurements based on transcriptions of actual aircraft signals. The subjective judgments made by human observers support the conclusion that the simulated sound closely approximates transcribed sounds of real aircraft.
Overview of an Indoor Sonic Boom Simulator at NASA Langley Research Center
NASA Technical Reports Server (NTRS)
Klos, Jacob
2012-01-01
A facility has been constructed at NASA Langley Research Center to simulate the soundscape inside residential houses that are exposed to environmental noise from aircraft. This controllable indoor listening environment, the Interior Effects Room, enables systematic study of parameters that affect psychoacoustic response. The single-room facility, built using typical residential construction methods and materials, is surrounded on adjacent sides by two arrays of loudspeakers in close proximity to the exterior walls. The arrays, containing 52 subwoofers and 52 mid-range speakers, have a usable bandwidth of 3 Hz to 5 kHz and sufficient output to allow study of sonic boom noise. In addition to these exterior arrays, satellite speakers placed inside the room are used to augment the transmitted sound with rattle and other audible contact ]induced noise that can result from low frequency excitation of a residential house. The layout of the facility, operational characteristics, acoustic characteristics and equalization approaches are summarized.
Assessment of auditory impression of the coolness and warmness of automotive HVAC noise.
Nakagawa, Seiji; Hotehama, Takuya; Kamiya, Masaru
2017-07-01
Noise induced by a heating, ventilation and air conditioning (HVAC) system in a vehicle is an important factor that affects the comfort of the interior of a car cabin. Much effort has been devoted to reduce noise levels, however, there is a need for a new sound design that addresses the noise problem from a different point of view. In this study, focusing on the auditory impression of automotive HVAC noise concerning coolness and warmness, psychoacoustical listening tests were performed using a paired comparison technique under various conditions of room temperature. Five stimuli were synthesized by stretching the spectral envelopes of recorded automotive HVAC noise to assess the effect of the spectral centroid, and were presented to normal-hearing subjects. Results show that the spectral centroid significantly affects the auditory impression concerning coolness and warmness; a higher spectral centroid induces a cooler auditory impression regardless of the room temperature.
Biomedical Simulation Models of Human Auditory Processes
NASA Technical Reports Server (NTRS)
Bicak, Mehmet M. A.
2012-01-01
Detailed acoustic engineering models that explore noise propagation mechanisms associated with noise attenuation and transmission paths created when using hearing protectors such as earplugs and headsets in high noise environments. Biomedical finite element (FE) models are developed based on volume Computed Tomography scan data which provides explicit external ear, ear canal, middle ear ossicular bones and cochlea geometry. Results from these studies have enabled a greater understanding of hearing protector to flesh dynamics as well as prioritizing noise propagation mechanisms. Prioritization of noise mechanisms can form an essential framework for exploration of new design principles and methods in both earplug and earcup applications. These models are currently being used in development of a novel hearing protection evaluation system that can provide experimentally correlated psychoacoustic noise attenuation. Moreover, these FE models can be used to simulate the effects of blast related impulse noise on human auditory mechanisms and brain tissue.
Behavioral and Molecular Genetics of Reading-Related AM and FM Detection Thresholds.
Bruni, Matthew; Flax, Judy F; Buyske, Steven; Shindhelm, Amber D; Witton, Caroline; Brzustowicz, Linda M; Bartlett, Christopher W
2017-03-01
Auditory detection thresholds for certain frequencies of both amplitude modulated (AM) and frequency modulated (FM) dynamic auditory stimuli are associated with reading in typically developing and dyslexic readers. We present the first behavioral and molecular genetic characterization of these two auditory traits. Two extant extended family datasets were given reading tasks and psychoacoustic tasks to determine FM 2 Hz and AM 20 Hz sensitivity thresholds. Univariate heritabilities were significant for both AM (h 2 = 0.20) and FM (h 2 = 0.29). Bayesian posterior probability of linkage (PPL) analysis found loci for AM (12q, PPL = 81 %) and FM (10p, PPL = 32 %; 20q, PPL = 65 %). Bivariate heritability analyses revealed that FM is genetically correlated with reading, while AM was not. Bivariate PPL analysis indicates that FM loci (10p, 20q) are not also associated with reading.
USAF bioenvironmental noise data handbook. Volume 157: KC-10A in-flight crew noise
NASA Astrophysics Data System (ADS)
Hille, H. K.
1982-09-01
The KC-10A is a standard USAF tanker-transport aircraft with high-speed, high altitude refueling and long range transport capability. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this helicopter during normal flight operations. Data are reported for 24 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Refer to Volume 1 of this handbook, USAF Bioenvironmental Noise Data Handbook, Vol. 1: Organization, Content and Application, AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
USAF Bioenvironmental Noise Data Handbook. Volume 165: MC-1 heater, duct type, portable
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-06-01
The MC-1 heater is a gasoline-motor driven, portable ground heater used primarily for cockpit and cabin temperature control. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at normal rated conditions. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
USAF Bioenvironmental Noise Data Handbook. Volume 152: C-12A in-flight crew noise
NASA Astrophysics Data System (ADS)
Hille, H. K.
1982-09-01
The C-12A is a military version of the Beechcraft Super King Air 200. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this aircraft during normal flight operations. Data are reported for five locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Refer to Volume 1 of this handbook, USAF Bioenvironmental Noise Data Handbook, Vol 1: Organization, Content and Application, AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
USAF Bioenvironmental Noise Data Handbook. Volume 156. HH-1N In-flight Crew Noise
NASA Astrophysics Data System (ADS)
Hille, H. K.
1982-11-01
The HH-IN is a USAF multi-purpose utility helicopter providing support for various USAF missions. This report provides measured data defining the bioacoustic environments at flight crew locations inside this helicopter during normal flight operations. Data are reported for two locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Refer to Volume 1 of this handbook, USAF Bioenvironmental Noise Data Handbook, Vol. 1: Organization, Content and Application, AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
USAF bioenvironmental noise data handbook. Volume 158: F-106A aircraft, near and far-field noise
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The USAF F-106A is a single seat, all-weather fighter/interceptor aircraft powered by a J75-P-17 turbojet engine. This report provides measured and extrapolated data defining the bioacoustic environments produced by this aircraft operating on a concrete runup pad for five engine-power conditions. Near-field data are reported for five locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 19 locations are normalized to standard meteorological conditions and extrapolated from 75 - 8000 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
USAF Bioenvironmental Noise Data Handbook, volume 154
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The E-4B is a Boeing 747 aircraft modified to serve as the national emergency/HQ Strategic Air Command Airborne Command Post. This report provides measured data defining the bioacoustic environments at flight crew/passenger locations inside this aircraft during normal flight operations. Data are reported for 24 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Refer to Volume 1 of this handbook, "USAF Bioenvironmental Noise Data Handbook, Vol. 1: Organization, Content and Application", AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
USAF bioenvironmental noise data handbook. Volume 163: GPC-28 compressor
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The GPC-28 is a gasoline engine-driven compressor with a 120 volt 60 Hz generator used for general purpose maintenance. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at a normal rated condition. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 - 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
USAF bioenvironmental noise data handbook. Volume 161: A/M32A-86 generator set, diesel engine driven
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The A/M32A-86 generator set is a diesel engine driven source of electrical power used for the starting of aircraft, and for ground maintenance. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at normal rated/loaded conditions. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 - 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
Pérez Zaballos, M T; Ramos Macías, Á; Pérez Placencia, D; Borkoski Barreiro, S A; Ramos de Miguel, Á
2016-06-01
Develop an online application that sends questions at different times of the day over an interval set by the user. It consists of a web interface that sends out WhatsApp messages containing questions from already validated or customized questionnaires (previous license) and shows the evolution of each question and patient over time and their average. The Laboratory of Psychoacoustics developed a platform that allows patients to answer surveys from remote locations. Hence, the developed tool allows to: (1) answer a test repeatedly under different personal situations and (2) evaluate the patient remotely. Hence, obtaining a more reliable assessment of the patient's QoL. Preliminary results using 5 teenagers aged 12-18 that answered the KINDL questionnaire is used to evidence the value of this new tool. Copyright © 2016 Elsevier Masson SAS. All rights reserved.
Development of a directivity-controlled piezoelectric transducer for sound reproduction
NASA Astrophysics Data System (ADS)
Bédard, Magella; Berry, Alain
2008-04-01
Present sound reproduction systems do not attempt to simulate the spatial radiation of musical instruments, or sound sources in general, even though the spatial directivity has a strong impact on the psychoacoustic experience. A transducer consisting of 4 piezoelectric elemental sources made from curved PVDF films is used to generate a target directivity pattern in the horizontal plane, in the frequency range of 5-20 kHz. The vibratory and acoustical response of an elemental source is addressed, both theoretically and experimentally. Two approaches to synthesize the input signals to apply to each elemental source are developed in order to create a prescribed, frequency-dependent acoustic directivity. The circumferential Fourier decomposition of the target directivity provides a compromise between the magnitude and the phase reconstruction, whereas the minimization of a quadratic error criterion provides a best magnitude reconstruction. This transducer can improve sound reproduction by introducing the spatial radiation aspect of the original source at high frequency.
NASA Technical Reports Server (NTRS)
Pearsons, Karl S.; Howe, Richard R.; Sneddon, Matthew D.; Fidell, Sanford
1996-01-01
Thirty audiometrically screened test participants judged the relative annoyance of two comparison (variable level) and thirty-four standard (fixed level) signals in an adaptive paired comparison psychoacoustic study. The signal ensemble included both FAR Part 36 Stage 2 and 3 aircraft overflights, as well as synthesized aircraft noise signatures and other non-aircraft signals. All test signals were presented for judgment as heard indoors, in the presence of continuous background noise, under free-field listening conditions in an anechoic chamber. Analyses of the performance of 30 noise metrics as predictors of these annoyance judgments confirmed that the more complex metrics were generally more accurate and precise predictors than the simpler methods. EPNL was somewhat less accurate and precise as a predictor of the annoyance judgments than a duration-adjusted variant of Zwicker's Loudness Level.
NASA Technical Reports Server (NTRS)
Hubbard, H. H.; Powell, C. A.
1981-01-01
A number of facilities were developed which provide a unique test capability for psychoacoustics and related human factors research. The design philosophy, physical layouts, dimensions, construction features, operating capabilities, and example applications for these facilities are described. In the exterior effects room, human subjects are exposed to the types of noises that are experienced outdoors, and in the interior effects room, subjects are exposed to the types of noises and noise-induced vibrations that are experience indoors. Subjects are also exposed to noises in an echo-free environment in the anechoic listening room. An aircraft noise synthesis system, which simulates aircraft flyover noise at an observer position on the ground, is used in conjunction with these three rooms. The passenger ride quality apparatus, a device for studying passenger response to noise and vibration in aircraft, or in other vehicles, is described.
Winn, Matthew B; Won, Jong Ho; Moon, Il Joon
This study was conducted to measure auditory perception by cochlear implant users in the spectral and temporal domains, using tests of either categorization (using speech-based cues) or discrimination (using conventional psychoacoustic tests). The authors hypothesized that traditional nonlinguistic tests assessing spectral and temporal auditory resolution would correspond to speech-based measures assessing specific aspects of phonetic categorization assumed to depend on spectral and temporal auditory resolution. The authors further hypothesized that speech-based categorization performance would ultimately be a superior predictor of speech recognition performance, because of the fundamental nature of speech recognition as categorization. Nineteen cochlear implant listeners and 10 listeners with normal hearing participated in a suite of tasks that included spectral ripple discrimination, temporal modulation detection, and syllable categorization, which was split into a spectral cue-based task (targeting the /ba/-/da/ contrast) and a timing cue-based task (targeting the /b/-/p/ and /d/-/t/ contrasts). Speech sounds were manipulated to contain specific spectral or temporal modulations (formant transitions or voice onset time, respectively) that could be categorized. Categorization responses were quantified using logistic regression to assess perceptual sensitivity to acoustic phonetic cues. Word recognition testing was also conducted for cochlear implant listeners. Cochlear implant users were generally less successful at utilizing both spectral and temporal cues for categorization compared with listeners with normal hearing. For the cochlear implant listener group, spectral ripple discrimination was significantly correlated with the categorization of formant transitions; both were correlated with better word recognition. Temporal modulation detection using 100- and 10-Hz-modulated noise was not correlated either with the cochlear implant subjects' categorization of voice onset time or with word recognition. Word recognition was correlated more closely with categorization of the controlled speech cues than with performance on the psychophysical discrimination tasks. When evaluating people with cochlear implants, controlled speech-based stimuli are feasible to use in tests of auditory cue categorization, to complement traditional measures of auditory discrimination. Stimuli based on specific speech cues correspond to counterpart nonlinguistic measures of discrimination, but potentially show better correspondence with speech perception more generally. The ubiquity of the spectral (formant transition) and temporal (voice onset time) stimulus dimensions across languages highlights the potential to use this testing approach even in cases where English is not the native language.
Winn, Matthew B.; Won, Jong Ho; Moon, Il Joon
2016-01-01
Objectives This study was conducted to measure auditory perception by cochlear implant users in the spectral and temporal domains, using tests of either categorization (using speech-based cues) or discrimination (using conventional psychoacoustic tests). We hypothesized that traditional nonlinguistic tests assessing spectral and temporal auditory resolution would correspond to speech-based measures assessing specific aspects of phonetic categorization assumed to depend on spectral and temporal auditory resolution. We further hypothesized that speech-based categorization performance would ultimately be a superior predictor of speech recognition performance, because of the fundamental nature of speech recognition as categorization. Design Nineteen CI listeners and 10 listeners with normal hearing (NH) participated in a suite of tasks that included spectral ripple discrimination (SRD), temporal modulation detection (TMD), and syllable categorization, which was split into a spectral-cue-based task (targeting the /ba/-/da/ contrast) and a timing-cue-based task (targeting the /b/-/p/ and /d/-/t/ contrasts). Speech sounds were manipulated in order to contain specific spectral or temporal modulations (formant transitions or voice onset time, respectively) that could be categorized. Categorization responses were quantified using logistic regression in order to assess perceptual sensitivity to acoustic phonetic cues. Word recognition testing was also conducted for CI listeners. Results CI users were generally less successful at utilizing both spectral and temporal cues for categorization compared to listeners with normal hearing. For the CI listener group, SRD was significantly correlated with the categorization of formant transitions; both were correlated with better word recognition. TMD using 100 Hz and 10 Hz modulated noise was not correlated with the CI subjects’ categorization of VOT, nor with word recognition. Word recognition was correlated more closely with categorization of the controlled speech cues than with performance on the psychophysical discrimination tasks. Conclusions When evaluating people with cochlear implants, controlled speech-based stimuli are feasible to use in tests of auditory cue categorization, to complement traditional measures of auditory discrimination. Stimuli based on specific speech cues correspond to counterpart non-linguistic measures of discrimination, but potentially show better correspondence with speech perception more generally. The ubiquity of the spectral (formant transition) and temporal (VOT) stimulus dimensions across languages highlights the potential to use this testing approach even in cases where English is not the native language. PMID:27438871
Acoustic fine structure may encode biologically relevant information for zebra finches.
Prior, Nora H; Smith, Edward; Lawson, Shelby; Ball, Gregory F; Dooling, Robert J
2018-04-18
The ability to discriminate changes in the fine structure of complex sounds is well developed in birds. However, the precise limit of this discrimination ability and how it is used in the context of natural communication remains unclear. Here we describe natural variability in acoustic fine structure of male and female zebra finch calls. Results from psychoacoustic experiments demonstrate that zebra finches are able to discriminate extremely small differences in fine structure, which are on the order of the variation in acoustic fine structure that is present in their vocal signals. Results from signal analysis methods also suggest that acoustic fine structure may carry information that distinguishes between biologically relevant categories including sex, call type and individual identity. Combined, our results are consistent with the hypothesis that zebra finches can encode biologically relevant information within the fine structure of their calls. This study provides a foundation for our understanding of how acoustic fine structure may be involved in animal communication.
Robinson, Philip W; Pätynen, Jukka; Lokki, Tapio; Jang, Hyung Suk; Jeon, Jin Yong; Xiang, Ning
2013-06-01
In musical or theatrical performance, some venues allow listeners to individually localize and segregate individual performers, while others produce a well blended ensemble sound. The room acoustic conditions that make this possible, and the psycho-acoustic effects at work are not fully understood. This research utilizes auralizations from measured and simulated performance venues to investigate spatial discrimination of multiple acoustic sources in rooms. Signals were generated from measurements taken in a small theater, and listeners in the audience area were asked to distinguish pairs of speech sources on stage with various spatial separations. This experiment was repeated with the proscenium splay walls treated to be flat, diffusive, or absorptive. Similar experiments were conducted in a simulated hall, utilizing 11 early reflections with various characteristics, and measured late reverberation. The experiments reveal that discriminating the lateral arrangement of two sources is possible at narrower separation angles when reflections come from flat or absorptive rather than diffusive surfaces.
Demodulation processes in auditory perception
NASA Astrophysics Data System (ADS)
Feth, Lawrence L.
1994-08-01
The long range goal of this project is the understanding of human auditory processing of information conveyed by complex, time-varying signals such as speech, music or important environmental sounds. Our work is guided by the assumption that human auditory communication is a 'modulation - demodulation' process. That is, we assume that sound sources produce a complex stream of sound pressure waves with information encoded as variations ( modulations) of the signal amplitude and frequency. The listeners task then is one of demodulation. Much of past. psychoacoustics work has been based in what we characterize as 'spectrum picture processing.' Complex sounds are Fourier analyzed to produce an amplitude-by-frequency 'picture' and the perception process is modeled as if the listener were analyzing the spectral picture. This approach leads to studies such as 'profile analysis' and the power-spectrum model of masking. Our approach leads us to investigate time-varying, complex sounds. We refer to them as dynamic signals and we have developed auditory signal processing models to help guide our experimental work.
A behavioral audiogram of the red fox (Vulpes vulpes).
Malkemper, E Pascal; Topinka, Václav; Burda, Hynek
2015-02-01
We determined the absolute hearing sensitivity of the red fox (Vulpes vulpes) using an adapted standard psychoacoustic procedure. The animals were tested in a reward-based go/no-go procedure in a semi-anechoic chamber. At 60 dB sound pressure level (SPL) (re 20 μPa) red foxes perceive pure tones between 51 Hz and 48 kHz, spanning 9.84 octaves with a single peak sensitivity of -15 dB at 4 kHz. The red foxes' high-frequency cutoff is comparable to that of the domestic dog while the low-frequency cutoff is comparable to that of the domestic cat and the absolute sensitivity is between both species. The maximal absolute sensitivity of the red fox is among the best found to date in any mammal. The procedure used here allows for assessment of animal auditory thresholds using positive reinforcement outside the laboratory. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Xing, Y. F.; Wang, Y. S.; Shi, L.; Guo, H.; Chen, H.
2016-01-01
According to the human perceptional characteristics, a method combined by the optimal wavelet-packet transform and artificial neural network, so-called OWPT-ANN model, for psychoacoustical recognition is presented. Comparisons of time-frequency analysis methods are performed, and an OWPT with 21 critical bands is designed for feature extraction of a sound, as is a three-layer back-propagation ANN for sound quality (SQ) recognition. Focusing on the loudness and sharpness, the OWPT-ANN model is applied on vehicle noises under different working conditions. Experimental verifications show that the OWPT can effectively transfer a sound into a time-varying energy pattern as that in the human auditory system. The errors of loudness and sharpness of vehicle noise from the OWPT-ANN are all less than 5%, which suggest a good accuracy of the OWPT-ANN model in SQ recognition. The proposed methodology might be regarded as a promising technique for signal processing in the human-hearing related fields in engineering.
The use of delay in multitrack production
NASA Astrophysics Data System (ADS)
Case, Alexander U.
2003-04-01
Delay, inevitable whenever sound propagates through space, is too often the bane of the acoustician's practice. An audible echo generally relegates a music performance hall-no matter how beautiful it otherwise might sound-to the lowest status. Multitrack music production on the other hand, with its aggressive use of overdubbing, editing, and signal processing, is not bound by those rules of time and space which determine the sound of a hall. In the recording studio, where music is synthesized for playback over loudspeakers, the delay is employed as a powerful, multipurpose tool. It is not avoided. It is in fact embraced. Echoes are used on purpose, strategically, to enhance the loudspeaker listening experience. Moreover, the humble delay is the basis for many nonecho effects. Flanging, chorus, and pitch shifting are delay-based effects regularly used in audio engineering practice. This paper discusses some of the more common delay-based effects, reviewing their technical structure, the psychoacoustic motivation behind them, and the musical value they create.
Pleasurable Emotional Response to Music: A Case of Neurodegenerative Generalized Auditory Agnosia
Matthews, Brandy R.; Chang, Chiung-Chih; De May, Mary; Engstrom, John; Miller, Bruce L.
2009-01-01
Recent functional neuroimaging studies implicate the network of mesolimbic structures known to be active in reward processing as the neural substrate of pleasure associated with listening to music. Psychoacoustic and lesion studies suggest that there is a widely distributed cortical network involved in processing discreet musical variables. Here we present the case of a young man with auditory agnosia as the consequence of cortical neurodegeneration who continues to experience pleasure when exposed to music. In a series of musical tasks the subject was unable to accurately identify any of the perceptual components of music beyond simple pitch discrimination, including musical variables know to impact the perception of affect. The subject subsequently misidentified the musical character of personally familiar tunes presented experimentally, but continued to report the activity of “listening” to specific musical genres was an emotionally rewarding experience. The implications of this case for the evolving understanding of music perception, music misperception, music memory, and music-associated emotion are discussed. PMID:19253088
Pleasurable emotional response to music: a case of neurodegenerative generalized auditory agnosia.
Matthews, Brandy R; Chang, Chiung-Chih; De May, Mary; Engstrom, John; Miller, Bruce L
2009-06-01
Recent functional neuroimaging studies implicate the network of mesolimbic structures known to be active in reward processing as the neural substrate of pleasure associated with listening to music. Psychoacoustic and lesion studies suggest that there is a widely distributed cortical network involved in processing discreet musical variables. Here we present the case of a young man with auditory agnosia as the consequence of cortical neurodegeneration who continues to experience pleasure when exposed to music. In a series of musical tasks, the subject was unable to accurately identify any of the perceptual components of music beyond simple pitch discrimination, including musical variables known to impact the perception of affect. The subject subsequently misidentified the musical character of personally familiar tunes presented experimentally, but continued to report that the activity of 'listening' to specific musical genres was an emotionally rewarding experience. The implications of this case for the evolving understanding of music perception, music misperception, music memory, and music-associated emotion are discussed.
It's about time: Presentation in honor of Ira Hirsh
NASA Astrophysics Data System (ADS)
Grant, Ken
2002-05-01
Over his long and illustrious career, Ira Hirsh has returned time and time again to his interest in the temporal aspects of pattern perception. Although Hirsh has studied and published articles and books pertaining to many aspects of the auditory system, such as sound conduction in the ear, cochlear mechanics, masking, auditory localization, psychoacoustic behavior in animals, speech perception, medical and audiological applications, coupling between psychophysics and physiology, and ecological acoustics, it is his work on auditory timing of simple and complex rhythmic patterns, the backbone of speech and music, that are at the heart of his more recent work. Here, we will focus on several aspects of temporal processing of simple and complex signals, both within and across sensory systems. Data will be reviewed on temporal order judgments of simple tones, and simultaneity judgments and intelligibility of unimodal and bimodal complex stimuli where stimulus components are presented either synchronously or asynchronously. Differences in the symmetry and shape of ``temporal windows'' derived from these data sets will be highlighted.
Acoustic Performance of a Real-Time Three-Dimensional Sound-Reproduction System
NASA Technical Reports Server (NTRS)
Faller, Kenneth J., II; Rizzi, Stephen A.; Aumann, Aric R.
2013-01-01
The Exterior Effects Room (EER) is a 39-seat auditorium at the NASA Langley Research Center and was built to support psychoacoustic studies of aircraft community noise. The EER has a real-time simulation environment which includes a three-dimensional sound-reproduction system. This system requires real-time application of equalization filters to compensate for spectral coloration of the sound reproduction due to installation and room effects. This paper describes the efforts taken to develop the equalization filters for use in the real-time sound-reproduction system and the subsequent analysis of the system s acoustic performance. The acoustic performance of the compensated and uncompensated sound-reproduction system is assessed for its crossover performance, its performance under stationary and dynamic conditions, the maximum spatialized sound pressure level it can produce from a single virtual source, and for the spatial uniformity of a generated sound field. Additionally, application examples are given to illustrate the compensated sound-reproduction system performance using recorded aircraft flyovers
Medium-Level Laser in Chronic Tinnitus Treatment
Dejakum, K.; Piegger, J.; Plewka, C.; Gunkel, A.; Thumfart, W.; Kudaibergenova, S.; Goebel, G.; Kral, F.; Freysinger, W.
2013-01-01
The purpose of this study was to evaluate the effect of medium-level laser therapy in chronic tinnitus treatment. In a prospective double-blind placebo-controlled trial, either active laser (450 mW, 830 nm combined Ga-Al-As diode laser) or placebo irradiation was applied through the external acoustic meatus of the affected ear towards the cochlea. Fourty-eight patients with chronic tinnitus were studied. The main outcome was measured using the Goebel tinnitus questionnaire, visual analogue scales measuring the perceived loudness of tinnitus, the annoyance associated with tinnitus, and the degree of attention paid to tinnitus as well as psycho-acoustical matches of tinnitus pitch and loudness. The results did show only very moderate temporary improvement of tinnitus. Moreover, no statistically relevant differences between laser and placebo group could be found. We conclude that medium-level laser therapy cannot be regarded as an effective treatment of chronic tinnitus in our therapy regime considering the limited number of patients included in our study. PMID:24294604
Medium-level laser in chronic tinnitus treatment.
Dejakum, K; Piegger, J; Plewka, C; Gunkel, A; Thumfart, W; Kudaibergenova, S; Goebel, G; Kral, F; Freysinger, W
2013-01-01
The purpose of this study was to evaluate the effect of medium-level laser therapy in chronic tinnitus treatment. In a prospective double-blind placebo-controlled trial, either active laser (450 mW, 830 nm combined Ga-Al-As diode laser) or placebo irradiation was applied through the external acoustic meatus of the affected ear towards the cochlea. Fourty-eight patients with chronic tinnitus were studied. The main outcome was measured using the Goebel tinnitus questionnaire, visual analogue scales measuring the perceived loudness of tinnitus, the annoyance associated with tinnitus, and the degree of attention paid to tinnitus as well as psycho-acoustical matches of tinnitus pitch and loudness. The results did show only very moderate temporary improvement of tinnitus. Moreover, no statistically relevant differences between laser and placebo group could be found. We conclude that medium-level laser therapy cannot be regarded as an effective treatment of chronic tinnitus in our therapy regime considering the limited number of patients included in our study.
Human annoyance and reactions to hotel room specific noises
NASA Astrophysics Data System (ADS)
Everhard, Ian L.
2004-05-01
A new formula is presented where multiple annoyance sources and transmission loss values of any partition are combined to produce a new single number rating of annoyance. The explanation of the formula is based on theoretical psychoacoustics and survey testing used to create variables used to weight the results. An imaginary hotel room is processed through the new formula and is rated based on theoretical survey results that would be taken by guests of the hotel. The new single number rating compares the multiple sources of annoyance to a single imaginary unbiased source where absolute level is the only factor in stimulating a linear rise in annoyance [Fidell et al., J. Acoust. Soc. Am. 66, 1427 (1979); D. M. Jones and D. E. Broadbent, ``Human performance and noise,'' in Handbook of Noise Control, 3rd ed., edited by C. M. Harris (ASA, New York, 1998), Chap. 24; J. P. Conroy and J. S. Roland, ``STC Field Testing and Results,'' in Sound and Vibration Magazine, Acoustical Publications, pp. 10-15 (July 2003)].
Evaluation of an Indoor Sonic Boom Subjective Test Facility at NASA Langley Research Center
NASA Technical Reports Server (NTRS)
Loubeau, Alexandra; Rathsam, Jonathan; Klos, Jacob
2011-01-01
A sonic boom simulator at NASA Langley Research Center has been constructed for research on human response to low-amplitude sonic booms heard indoors. Research in this facility will ultimately lead to development of a psychoacoustic model for single indoor booms. The first subjective test was designed to explore indoor human response to variations in sonic boom rise time and amplitude. Another goal was to identify loudness level variability across listener locations within the facility. Finally, the test also served to evaluate the facility as a laboratory research tool for studying indoor human response to sonic booms. Subjects listened to test sounds and were asked to rate their annoyance relative to a reference boom. Measurements of test signals were conducted for objective analysis and correlation with subjective responses. Results confirm the functionality of the facility and effectiveness of the test methods and indicate that loudness level does not fully describe indoor annoyance to the selected sonic boom signals.
Sensitivity to musical emotion is influenced by tonal structure in congenital amusia.
Jiang, Cunmei; Liu, Fang; Wong, Patrick C M
2017-08-08
Emotional communication in music depends on multiple attributes including psychoacoustic features and tonal system information, the latter of which is unique to music. The present study investigated whether congenital amusia, a lifelong disorder of musical processing, impacts sensitivity to musical emotion elicited by timbre and tonal system information. Twenty-six amusics and 26 matched controls made tension judgments on Western (familiar) and Indian (unfamiliar) melodies played on piano and sitar. Like controls, amusics used timbre cues to judge musical tension in Western and Indian melodies. While controls assigned significantly lower tension ratings to Western melodies compared to Indian melodies, thus showing a tonal familiarity effect on tension ratings, amusics provided comparable tension ratings for Western and Indian melodies on both timbres. Furthermore, amusics rated Western melodies as more tense compared to controls, as they relied less on tonality cues than controls in rating tension for Western melodies. The implications of these findings in terms of emotional responses to music are discussed.
The Efficacy of Auditory Perceptual Training for Tinnitus: A Systematic Review
Stacey, Paula C.; Hall, Deborah A.
2010-01-01
Auditory perceptual training affects neural plasticity and so represents a potential strategy for tinnitus management. We assessed the effects of auditory perceptual training on tinnitus perception and/or its intrusiveness via a systematic review of published literature. An electronic database search using the keywords ‘tinnitus and learning’ or ‘tinnitus and training’ was conducted, updated by a hand search. The ten studies identified were reviewed independently by two reviewers, data were extracted, study quality was assessed according to a number of specific criteria and the information was synthesised using a narrative approach. Nine out of the ten studies reported some significant change in either self-reported or psychoacoustic outcome measures after auditory training. However, all studies were quality rated as providing low or moderate levels of evidence for an effect. We identify a need for appropriately randomised and controlled studies that will generate high-quality unbiased and generalisable evidence to ascertain whether or not auditory perceptual training has a clinically relevant effect on tinnitus. PMID:20668974
Feasibility study of a game integrating assessment and therapy of tinnitus.
Wise, K; Kobayashi, K; Searchfield, G D
2015-07-15
Tinnitus, head and ear noise, is due to maladaptive plastic changes in auditory and associated neural networks. Tinnitus has been traditionally managed through the use of sound to passively mask or facilitate habituation to tinnitus, a process that may take 6-12 months. A game-based perceptual training method, requiring localisation and selective attention to sounds, was developed and customised to the individual's tinnitus perception. Eight participants tested the games usability at home. Each participant successfully completed 30 min of training, for 20 days, along with daily psychoacoustic assessment of tinnitus pitch and loudness. The training period and intensity of training appears sufficient to reduce tinnitus handicap. The training approach used may be a viable alternative to frequency discrimination based training for treating tinnitus (Hoare et al., 2014) and a useful tool in exploring learning mechanisms in the auditory system. Integration of tinnitus assessment with therapy in a game is feasible, and the method(s) warrant further investigation. Copyright © 2015 Elsevier B.V. All rights reserved.
Won, Jong Ho; Jones, Gary L; Drennan, Ward R; Jameyson, Elyse M; Rubinstein, Jay T
2011-10-01
Spectral-ripple discrimination has been used widely for psychoacoustical studies in normal-hearing, hearing-impaired, and cochlear implant listeners. The present study investigated the perceptual mechanism for spectral-ripple discrimination in cochlear implant listeners. The main goal of this study was to determine whether cochlear implant listeners use a local intensity cue or global spectral shape for spectral-ripple discrimination. The effect of electrode separation on spectral-ripple discrimination was also evaluated. Results showed that it is highly unlikely that cochlear implant listeners depend on a local intensity cue for spectral-ripple discrimination. A phenomenological model of spectral-ripple discrimination, as an "ideal observer," showed that a perceptual mechanism based on discrimination of a single intensity difference cannot account for performance of cochlear implant listeners. Spectral modulation depth and electrode separation were found to significantly affect spectral-ripple discrimination. The evidence supports the hypothesis that spectral-ripple discrimination involves integrating information from multiple channels. © 2011 Acoustical Society of America
Ho Won, Jong; Jones, Gary L.; Drennan, Ward R.; Jameyson, Elyse M.; Rubinstein, Jay T.
2011-01-01
Spectral-ripple discrimination has been used widely for psychoacoustical studies in normal-hearing, hearing-impaired, and cochlear implant listeners. The present study investigated the perceptual mechanism for spectral-ripple discrimination in cochlear implant listeners. The main goal of this study was to determine whether cochlear implant listeners use a local intensity cue or global spectral shape for spectral-ripple discrimination. The effect of electrode separation on spectral-ripple discrimination was also evaluated. Results showed that it is highly unlikely that cochlear implant listeners depend on a local intensity cue for spectral-ripple discrimination. A phenomenological model of spectral-ripple discrimination, as an “ideal observer,” showed that a perceptual mechanism based on discrimination of a single intensity difference cannot account for performance of cochlear implant listeners. Spectral modulation depth and electrode separation were found to significantly affect spectral-ripple discrimination. The evidence supports the hypothesis that spectral-ripple discrimination involves integrating information from multiple channels. PMID:21973363
A Summary of NASA Research Exploring the Acoustics of Small Unmanned Aerial Systems
NASA Technical Reports Server (NTRS)
Zawodny, Nikolas S.; Christian, Andrew; Cabell, Randolph
2018-01-01
Proposed uses of small unmanned aerial systems (sUAS) have the potential to expose large portions of communities to a new noise source. In order to understand the potential noise impact of sUAS, NASA initiated acoustics research as one component of the 3-year DELIVER project, with the goal of documenting the feasibility of using existing aircraft design tools and methods on this class of vehicles. This paper summarizes the acoustics research conducted within the DELIVER project. The research described here represents an initial study, and subsequent research building on the findings of this work has been proposed for other NASA projects. The paper summarizes acoustics research in four areas: measurements of noise generated by flyovers of small unmanned aerial vehicles, measurements in controlled test facilities to understand the noise generated by components of these vehicles, computational predictions of component and full vehicle noise, and psychoacoustic tests including auralizations conducted to assess human annoyance to the noise generated by these vehicles.
Perceived Annoyance to Noise Produced by a Distributed Electric Propulsion High Lift System
NASA Technical Reports Server (NTRS)
Palumbo, Dan; Rathsam, Jonathan; Christian, Andrew; Rafaelof, Menachem
2016-01-01
Results of a psychoacoustic test performed to understand the relative annoyance to noise produced by several configurations of a distributed electric propulsion high lift system are given. It is found that the number of propellers in the system is a major factor in annoyance perception. This is an intuitive result as annoyance increases, in general, with frequency, and, the blade passage frequency of the propellers increases with the number of propellers. Additionally, the data indicate that having some variation in the blade passage frequency from propeller-to-propeller is beneficial as it reduces the high tonality generated when all the propellers are spinning in synchrony at the same speed. The propellers can be set to spin at different speeds, but it was found that allowing the motor controllers to drift within 1% of nominal settings produced the best results (lowest overall annoyance). The methodology employed has been demonstrated to be effective in providing timely feedback to designers in the early stages of design development.
Mendez, M F
2001-02-01
After a right temporoparietal stroke, a left-handed man lost the ability to understand speech and environmental sounds but developed greater appreciation for music. The patient had preserved reading and writing but poor verbal comprehension. Slower speech, single syllable words, and minimal written cues greatly facilitated his verbal comprehension. On identifying environmental sounds, he made predominant acoustic errors. Although he failed to name melodies, he could match, describe, and sing them. The patient had normal hearing except for presbyacusis, right-ear dominance for phonemes, and normal discrimination of basic psychoacoustic features and rhythm. Further testing disclosed difficulty distinguishing tone sequences and discriminating two clicks and short-versus-long tones, particularly in the left ear. Together, these findings suggest impairment in a direct route for temporal analysis and auditory word forms in his right hemisphere to Wernicke's area in his left hemisphere. The findings further suggest a separate and possibly rhythm-based mechanism for music recognition.
Tinnitus severity, depression, and the big five personality traits.
Langguth, B; Kleinjung, T; Fischer, B; Hajak, G; Eichhammer, P; Sand, P G
2007-01-01
A growing number of self-report measures for the evaluation of tinnitus severity has become available to research and clinical practice. This has led to an increased awareness of depression and personality as predictors of tinnitus severity in addition to loudness and other psychoacoustic measures. However, the net impact of personality dimensions on tinnitus ratings has not been investigated when the effect of depressed mood is controlled. In the present study, we demonstrate the role of the big five personality traits, 'Neuroticism', 'Extraversion', 'Openness', 'Agreeableness', and 'Conscientiousness', in affecting scores on two standard instruments for grading tinnitus-related complaints, the tinnitus handicap inventory (THI), and the tinnitus questionnaire (TQ). When 72 individuals with chronic tinnitus were examined, 'Agreeableness' negatively correlated with THI scores (p=.003), whereas the anxiety trait 'Neuroticism' correlated both with depressive symptomatology (p<.001) and TQ scores (p=.028), but not with THI ratings (n.s.). In addition to confirming the established roles of trait anxiety and depression, low 'Agreeableness' was thus identified as a novel predictor of tinnitus severity on the THI.
The Nature of Exhibits About Acoustics in Science and Technology Centres
NASA Astrophysics Data System (ADS)
Afonso, Ana S.; Gilbert, John K.
2008-11-01
This is a study of the opportunities currently provided by interactive science and technology centres for visitors’ engagement in the field of acoustics. E-mails, requesting a description of exhibits on acoustics (sound and hearing) in use, were sent to members of staff of interactive science and technology centres around the world as well as to companies that design and sell exhibits. Eighty-seven descriptions of distinctive interactive exhibits were received and analysed. Results show that: there are few analogy-based exhibits concerning the more complex aspects of acoustics; narratives involving visitors’ everyday lives, that might provide continuity between and beyond the situations presented by exhibits, are not generally provided; science is emphasised at the expense of technology; the risks, benefits and ethical implications of relevant technological artefacts are rarely mentioned; the majority of the exhibits are concerned with the fields of fundamental acoustics, hearing, and psychoacoustics. It is suggested that interactive science and technology centres need to rethink the design of exhibits about acoustics if their mission includes some appreciation of this important branch of science and technology.
Measuring the critical band for speech.
Healy, Eric W; Bacon, Sid P
2006-02-01
The current experiments were designed to measure the frequency resolution employed by listeners during the perception of everyday sentences. Speech bands having nearly vertical filter slopes and narrow bandwidths were sharply partitioned into various numbers of equal log- or ERBN-width subbands. The temporal envelope from each partition was used to amplitude modulate a corresponding band of low-noise noise, and the modulated carriers were combined and presented to normal-hearing listeners. Intelligibility increased and reached asymptote as the number of partitions increased. In the mid- and high-frequency regions of the speech spectrum, the partition bandwidth corresponding to asymptotic performance matched current estimates of psychophysical tuning across a number of conditions. These results indicate that, in these regions, the critical band for speech matches the critical band measured using traditional psychoacoustic methods and nonspeech stimuli. However, in the low-frequency region, partition bandwidths at asymptote were somewhat narrower than would be predicted based upon psychophysical tuning. It is concluded that, overall, current estimates of psychophysical tuning represent reasonably well the ability of listeners to extract spectral detail from running speech.
Salicylate-Induced Hearing Loss and Gap Detection Deficits in Rats
Radziwon, Kelly E.; Stolzberg, Daniel J.; Urban, Maxwell E.; Bowler, Rachael A.; Salvi, Richard J.
2015-01-01
To test the “tinnitus gap-filling” hypothesis in an animal psychoacoustic paradigm, rats were tested using a go/no-go operant gap detection task in which silent intervals of various durations were embedded within a continuous noise. Gap detection thresholds were measured before and after treatment with a dose of sodium salicylate (200 mg/kg) that reliably induces tinnitus in rats. Noise-burst detection thresholds were also measured to document the amount of hearing loss and aid in interpreting the gap detection results. As in the previous human psychophysical experiments, salicylate had little or no effect on gap thresholds measured in broadband noise presented at high-stimulus levels (30–60 dB SPL); gap detection thresholds were always 10 ms or less. Salicylate also did not affect gap thresholds presented in narrowband noise at 60 dB SPL. Therefore, rats treated with a dose of salicylate that reliably induces tinnitus have no difficulty detecting silent gaps as long as the noise in which they are embedded is clearly audible. PMID:25750635
Tong, Xiuhong; Tong, Xiuli; King Yiu, Fung
Increasing evidence suggests that children with developmental dyslexia exhibit a deficit not only at the segmental level of phonological processing but also, by extension, at the suprasegmental level. However, it remains unclear whether such a suprasegmental phonological processing deficit is due to a difficulty in processing acoustic cues of speech rhythm, such as rise time and intensity. This study set out to investigate to what extent suprasegmental phonological processing (i.e., Cantonese lexical tone perception) and rise time sensitivity could distinguish Chinese children with dyslexia from typically developing children. Sixteen children with dyslexia and 44 age-matched controls were administered a Cantonese lexical tone perception task, psychoacoustic tasks, a nonverbal reasoning ability task, and word reading and dictation tasks. Children with dyslexia performed worse than controls on Cantonese lexical tone perception, rise time, and intensity. Furthermore, Cantonese lexical tone perception appeared to be a stable indicator that distinguishes children with dyslexia from controls, even after controlling for basic auditory processing skills. These findings suggest that suprasegmental phonological processing (i.e., lexical tone perception) is a potential factor that accounts for reading difficulty in Chinese.
Audiogram of a striped dolphin (Stenella coeruleoalba)
NASA Astrophysics Data System (ADS)
Kastelein, Ronald A.; Hagedoorn, Monique; Au, Whitlow W. L.; de Haan, Dick
2003-02-01
The underwater hearing sensitivity of a striped dolphin was measured in a pool using standard psycho-acoustic techniques. The go/no-go response paradigm and up-down staircase psychometric method were used. Auditory sensitivity was measured by using 12 narrow-band frequency-modulated signals having center frequencies between 0.5 and 160 kHz. The 50% detection threshold was determined for each frequency. The resulting audiogram for this animal was U-shaped, with hearing capabilities from 0.5 to 160 kHz (8
Quinine reduces the dynamic range of the human auditory system.
Berninger, E; Karlsson, K K; Alván, G
1998-01-01
The aim of the study was to evaluate and quantify quinine-induced changes in the human auditory dynamic range, as a model for cochlear hearing loss. Six otologically normal volunteers (21-40 years old) received quinine hydrochloride (15 mg/kg body weight) in two identical oral doses and one intravenous infusion. Refined hearing tests were performed monaurally at threshold, at moderate hearing levels and at high hearing levels. Quinine induced a maximal pure-tone threshold shift of 23 dB (1000-2000 Hz). The increase in the psychoacoustical click threshold agreed with an increase in the detection threshold of click-evoked otoacoustic emissions. The change in the stimulus-response relationship of the emissions reflected recruitment. The self-attained most comfortable speech level and the acoustic stapedius reflex thresholds were not affected by quinine administration. Quinine is a useful model substance for reversibly inducing complete loudness recruitment in humans as it acts specifically on some parts of the hearing function. Its mechanism of action on the molecular level is likely to reveal further information on the physiology of hearing.
Lee, Hyun-Ho; Lee, Sang-Kwon
2009-09-01
Booming sound is one of the important sounds in a passenger car. The aim of the paper is to develop the objective evaluation method of interior booming sound. The development method is based on the sound metrics and ANN (artificial neural network). The developed method is called the booming index. Previous work maintained that booming sound quality is related to loudness and sharpness--the sound metrics used in psychoacoustics--and that the booming index is developed by using the loudness and sharpness for a signal within whole frequency between 20 Hz and 20 kHz. In the present paper, the booming sound quality was found to be effectively related to the loudness at frequencies below 200 Hz; thus the booming index is updated by using the loudness of the signal filtered by the low pass filter at frequency under 200 Hz. The relationship between the booming index and sound metric is identified by an ANN. The updated booming index has been successfully applied to the objective evaluation of the booming sound quality of mass-produced passenger cars.
Spectral envelope sensitivity of musical instrument sounds.
Gunawan, David; Sen, D
2008-01-01
It is well known that the spectral envelope is a perceptually salient attribute in musical instrument timbre perception. While a number of studies have explored discrimination thresholds for changes to the spectral envelope, the question of how sensitivity varies as a function of center frequency and bandwidth for musical instruments has yet to be addressed. In this paper a two-alternative forced-choice experiment was conducted to observe perceptual sensitivity to modifications made on trumpet, clarinet and viola sounds. The experiment involved attenuating 14 frequency bands for each instrument in order to determine discrimination thresholds as a function of center frequency and bandwidth. The results indicate that perceptual sensitivity is governed by the first few harmonics and sensitivity does not improve when extending the bandwidth any higher. However, sensitivity was found to decrease if changes were made only to the higher frequencies and continued to decrease as the distorted bandwidth was widened. The results are analyzed and discussed with respect to two other spectral envelope discrimination studies in the literature as well as what is predicted from a psychoacoustic model.
Complex pitch perception mechanisms are shared by humans and a New World monkey.
Song, Xindong; Osmanski, Michael S; Guo, Yueqi; Wang, Xiaoqin
2016-01-19
The perception of the pitch of harmonic complex sounds is a crucial function of human audition, especially in music and speech processing. Whether the underlying mechanisms of pitch perception are unique to humans, however, is unknown. Based on estimates of frequency resolution at the level of the auditory periphery, psychoacoustic studies in humans have revealed several primary features of central pitch mechanisms. It has been shown that (i) pitch strength of a harmonic tone is dominated by resolved harmonics; (ii) pitch of resolved harmonics is sensitive to the quality of spectral harmonicity; and (iii) pitch of unresolved harmonics is sensitive to the salience of temporal envelope cues. Here we show, for a standard musical tuning fundamental frequency of 440 Hz, that the common marmoset (Callithrix jacchus), a New World monkey with a hearing range similar to that of humans, exhibits all of the primary features of central pitch mechanisms demonstrated in humans. Thus, marmosets and humans may share similar pitch perception mechanisms, suggesting that these mechanisms may have emerged early in primate evolution.
The effects of sad prosody on hemispheric specialization for words processing.
Leshem, Rotem; Arzouan, Yossi; Armony-Sivan, Rinat
2015-06-01
This study examined the effect of sad prosody on hemispheric specialization for word processing using behavioral and electrophysiological measures. A dichotic listening task combining focused attention and signal-detection methods was conducted to evaluate the detection of a word spoken in neutral or sad prosody. An overall right ear advantage together with leftward lateralization in early (150-170 ms) and late (240-260 ms) processing stages was found for word detection, regardless of prosody. Furthermore, the early stage was most pronounced for words spoken in neutral prosody, showing greater negative activation over the left than the right hemisphere. In contrast, the later stage was most pronounced for words spoken with sad prosody, showing greater positive activation over the left than the right hemisphere. The findings suggest that sad prosody alone was not sufficient to modulate hemispheric asymmetry in word-level processing. We posit that lateralized effects of sad prosody on word processing are largely dependent on the psychoacoustic features of the stimuli as well as on task demands. Copyright © 2015 Elsevier Inc. All rights reserved.
Tutorial on the Psychophysics and Technology of Virtual Acoustic Displays
NASA Technical Reports Server (NTRS)
Wenzel, Elizabeth M.; Null, Cynthia (Technical Monitor)
1998-01-01
Virtual acoustics, also known as 3-D sound and auralization, is the simulation of the complex acoustic field experienced by a listener within an environment. Going beyond the simple intensity panning of normal stereo techniques, the goal is to process sounds so that they appear to come from particular locations in three-dimensional space. Although loudspeaker systems are being developed, most of the recent work focuses on using headphones for playback and is the outgrowth of earlier analog techniques. For example, in binaural recording, the sound of an orchestra playing classical music is recorded through small mics in the two "ear canals" of an anthropomorphic artificial or "dummy" head placed in the audience of a concert hall. When the recorded piece is played back over headphones, the listener passively experiences the illusion of hearing the violins on the left and the cellos on the right, along with all the associated echoes, resonances, and ambience of the original environment. Current techniques use digital signal processing to synthesize the acoustical properties that people use to localize a sound source in space. Thus, they provide the flexibility of a kind of digital dummy head, allowing a more active experience in which a listener can both design and move around or interact with a simulated acoustic environment in real time. Such simulations are being developed for a variety of application areas including architectural acoustics, advanced human-computer interfaces, telepresence and virtual reality, navigation aids for the visually-impaired, and as a test bed for psychoacoustical investigations of complex spatial cues. The tutorial will review the basic psychoacoustical cues that determine human sound localization and the techniques used to measure these cues as Head-Related Transfer Functions (HRTFs) for the purpose of synthesizing virtual acoustic environments. The only conclusive test of the adequacy of such simulations is an operational one in which the localization of real and synthesized stimuli are directly compared in psychophysical studies. To this end, the results of psychophysical experiments examining the perceptual validity of the synthesis technique will be reviewed and factors that can enhance perceptual accuracy and realism will be discussed. Of particular interest is the relationship between individual differences in HRTFs and in behavior, the role of reverberant cues in reducing the perceptual errors observed with virtual sound sources, and the importance of developing perceptually valid methods of simplifying the synthesis technique. Recent attempts to implement the synthesis technique in real time systems will also be discussed and an attempt made to interpret their quoted system specifications in terms of perceptual performance. Finally, some critical research and technology development issues for the future will be outlined.
Shao, Yu; Chang, Chip-Hong
2007-08-01
We present a new speech enhancement scheme for a single-microphone system to meet the demand for quality noise reduction algorithms capable of operating at a very low signal-to-noise ratio. A psychoacoustic model is incorporated into the generalized perceptual wavelet denoising method to reduce the residual noise and improve the intelligibility of speech. The proposed method is a generalized time-frequency subtraction algorithm, which advantageously exploits the wavelet multirate signal representation to preserve the critical transient information. Simultaneous masking and temporal masking of the human auditory system are modeled by the perceptual wavelet packet transform via the frequency and temporal localization of speech components. The wavelet coefficients are used to calculate the Bark spreading energy and temporal spreading energy, from which a time-frequency masking threshold is deduced to adaptively adjust the subtraction parameters of the proposed method. An unvoiced speech enhancement algorithm is also integrated into the system to improve the intelligibility of speech. Through rigorous objective and subjective evaluations, it is shown that the proposed speech enhancement system is capable of reducing noise with little speech degradation in adverse noise environments and the overall performance is superior to several competitive methods.
ANALOGY AND DISANALOGY IN PRODUCTION AND PERCEPTION OF SPEECH
Remez, Robert E.
2014-01-01
A varied psychological vocabulary now describes the cognitive and social conditions of language production, the ultimate result of which is the mechanical action of vocal musculature in spoken expression. Following the logic of the speech chain, descriptions of production have often exhibited a clear analogy to accounts of perception. This reciprocality is especially evident in explanations that rely on reafference to control production, on articulation to inform perception, and on strict parity between produced and perceived form to provide invariance in the relation between abstract linguistic objects and observed expression. However, a causal account of production and perception cannot derive solely from this hopeful analogy. Despite sharing of abstract linguistic representations, the control functions in production and perception as well as the constraints on their use stand in fundamental disanalogy. This is readily seen in the different adaptive challenges to production — to speak in a single voice — and perception — to resolve familiar linguistic properties in any voice. This acknowledgment sets descriptive and theoretical challenges that break the symmetry of production and perception. As a consequence, this recognition dislodges an old impasse between the psychoacoustic and motoric accounts in the regulation of production and perception. PMID:25642428
NASA Astrophysics Data System (ADS)
Martens, William
2005-04-01
Several attributes of auditory spatial imagery associated with stereophonic sound reproduction are strongly modulated by variation in interaural cross correlation (IACC) within low frequency bands. Nonetheless, a standard practice in bass management for two-channel and multichannel loudspeaker reproduction is to mix low-frequency musical content to a single channel for reproduction via a single driver (e.g., a subwoofer). This paper reviews the results of psychoacoustic studies which support the conclusion that reproduction via multiple drivers of decorrelated low-frequency signals significantly affects such important spatial attributes as auditory source width (ASW), auditory source distance (ASD), and listener envelopment (LEV). A variety of methods have been employed in these tests, including forced choice discrimination and identification, and direct ratings of both global dissimilarity and distinct attributes. Contrary to assumptions that underlie industrial standards established in 1994 by ITU-R. Recommendation BS.775-1, these findings imply that substantial stereophonic spatial information exists within audio signals at frequencies below the 80 to 120 Hz range of prescribed subwoofer cutoff frequencies, and that loudspeaker reproduction of decorrelated signals at frequencies as low as 50 Hz can have an impact upon auditory spatial imagery. [Work supported by VRQ.
Rapid estimation of high-parameter auditory-filter shapes
Shen, Yi; Sivakumar, Rajeswari; Richards, Virginia M.
2014-01-01
A Bayesian adaptive procedure, the quick-auditory-filter (qAF) procedure, was used to estimate auditory-filter shapes that were asymmetric about their peaks. In three experiments, listeners who were naive to psychoacoustic experiments detected a fixed-level, pure-tone target presented with a spectrally notched noise masker. The qAF procedure adaptively manipulated the masker spectrum level and the position of the masker notch, which was optimized for the efficient estimation of the five parameters of an auditory-filter model. Experiment I demonstrated that the qAF procedure provided a convergent estimate of the auditory-filter shape at 2 kHz within 150 to 200 trials (approximately 15 min to complete) and, for a majority of listeners, excellent test-retest reliability. In experiment II, asymmetric auditory filters were estimated for target frequencies of 1 and 4 kHz and target levels of 30 and 50 dB sound pressure level. The estimated filter shapes were generally consistent with published norms, especially at the low target level. It is known that the auditory-filter estimates are narrower for forward masking than simultaneous masking due to peripheral suppression, a result replicated in experiment III using fewer than 200 qAF trials. PMID:25324086
Perception of intonation in Mandarin Chinese.
Yuan, Jiahong
2011-12-01
There is a tendency across languages to use a rising pitch contour to convey question intonation and a falling pitch contour to convey a statement. In a lexical tone language such as Mandarin Chinese, rising and falling pitch contours are also used to differentiate lexical meaning. How, then, does the multiplexing of the F(0) channel affect the perception of question and statement intonation in a lexical tone language? This study investigated the effects of lexical tones and focus on the perception of intonation in Mandarin Chinese. The results show that lexical tones and focus impact the perception of sentence intonation. Question intonation was easier for native speakers to identify on a sentence with a final falling tone and more difficult to identify on a sentence with a final rising tone, suggesting that tone identification intervenes in the mapping of F(0) contours to intonational categories and that tone and intonation interact at the phonological level. In contrast, there is no evidence that the interaction between focus and intonation goes beyond the psychoacoustic level. The results provide insights that will be useful for further research on tone and intonation interactions in both acoustic modeling studies and neurobiological studies. © 2011 Acoustical Society of America
Dietz, Mathias; Marquardt, Torsten; Salminen, Nelli H.; McAlpine, David
2013-01-01
The ability to locate the direction of a target sound in a background of competing sources is critical to the survival of many species and important for human communication. Nevertheless, brain mechanisms that provide for such accurate localization abilities remain poorly understood. In particular, it remains unclear how the auditory brain is able to extract reliable spatial information directly from the source when competing sounds and reflections dominate all but the earliest moments of the sound wave reaching each ear. We developed a stimulus mimicking the mutual relationship of sound amplitude and binaural cues, characteristic to reverberant speech. This stimulus, named amplitude modulated binaural beat, allows for a parametric and isolated change of modulation frequency and phase relations. Employing magnetoencephalography and psychoacoustics it is demonstrated that the auditory brain uses binaural information in the stimulus fine structure only during the rising portion of each modulation cycle, rendering spatial information recoverable in an otherwise unlocalizable sound. The data suggest that amplitude modulation provides a means of “glimpsing” low-frequency spatial cues in a manner that benefits listening in noisy or reverberant environments. PMID:23980161
Jürgens, Tim; Brand, Thomas
2009-11-01
This study compares the phoneme recognition performance in speech-shaped noise of a microscopic model for speech recognition with the performance of normal-hearing listeners. "Microscopic" is defined in terms of this model twofold. First, the speech recognition rate is predicted on a phoneme-by-phoneme basis. Second, microscopic modeling means that the signal waveforms to be recognized are processed by mimicking elementary parts of human's auditory processing. The model is based on an approach by Holube and Kollmeier [J. Acoust. Soc. Am. 100, 1703-1716 (1996)] and consists of a psychoacoustically and physiologically motivated preprocessing and a simple dynamic-time-warp speech recognizer. The model is evaluated while presenting nonsense speech in a closed-set paradigm. Averaged phoneme recognition rates, specific phoneme recognition rates, and phoneme confusions are analyzed. The influence of different perceptual distance measures and of the model's a-priori knowledge is investigated. The results show that human performance can be predicted by this model using an optimal detector, i.e., identical speech waveforms for both training of the recognizer and testing. The best model performance is yielded by distance measures which focus mainly on small perceptual distances and neglect outliers.
Dolphin biosonar target detection in noise: wrap up of a past experiment.
Au, Whitlow W L
2014-07-01
The target detection capability of bottlenose dolphins in the presence of artificial masking noise was first studied by Au and Penner [J. Acoust. Soc. Am. 70, 687-693 (1981)] in which the dolphins' target detection threshold was determined as a function of the ratio of the echo energy flux density and the estimated received noise spectral density. Such a metric was commonly used in human psychoacoustics despite the fact that the echo energy flux density is not compatible with noise spectral density which is averaged intensity per Hz. Since the earlier detection in noise studies, two important parameters, the dolphin integration time applicable to broadband clicks and the dolphin's auditory filter shape, were determined. The inclusion of these two parameters allows for the estimation of the received energy flux density of the masking noise so that the dolphin target detection can now be determined as a function of the ratio of the received energy of the echo over the received noise energy. Using an integration time of 264 μs and an auditory bandwidth of 16.7 kHz, the ratio of the echo energy to noise energy at the target detection threshold is approximately 1 dB.
NASA Astrophysics Data System (ADS)
Lauter, Judith
2002-05-01
As Research Director of CID, Ira emphasized the importance of combining information from biology with rigorous studies of behavior, such as psychophysics, to better understand how the brain and body accomplish the goals of everyday life. In line with this philosophy, my doctoral dissertation sought to explain brain functional asymmetries (studied with dichotic listening) in terms of the physical dimensions of a library of test sounds designed to represent a speech-music continuum. Results highlighted individual differences plus similarities in terms of patterns of relative ear advantages, suggesting an organizational basis for brain asymmetries depending on physical dimensions of stimulus and gesture with analogs in auditory, visual, somatosensory, and motor systems. My subsequent work has employed a number of noninvasive methods (OAEs, EPs, qEEG, PET, MRI) to explore the neurobiological bases of individual differences in general and functional asymmetries in particular. This research has led to (1) the AXS test battery for assessing the neurobiology of human sensory-motor function; (2) the handshaking model of brain function, describing dynamic relations along all three body/brain axes; (3) the four-domain EPIC model of functional asymmetries; and (4) the trimodal brain, a new model of individual differences based on psychoimmunoneuroendocrinology.
Acoustic Calibration of the Exterior Effects Room at the NASA Langley Research Center
NASA Technical Reports Server (NTRS)
Faller, Kenneth J., II; Rizzi, Stephen A.; Klos, Jacob; Chapin, William L.; Surucu, Fahri; Aumann, Aric R.
2010-01-01
The Exterior Effects Room (EER) at the NASA Langley Research Center is a 39-seat auditorium built for psychoacoustic studies of aircraft community noise. The original reproduction system employed monaural playback and hence lacked sound localization capability. In an effort to more closely recreate field test conditions, a significant upgrade was undertaken to allow simulation of a three-dimensional audio and visual environment. The 3D audio system consists of 27 mid and high frequency satellite speakers and 4 subwoofers, driven by a real-time audio server running an implementation of Vector Base Amplitude Panning. The audio server is part of a larger simulation system, which controls the audio and visual presentation of recorded and synthesized aircraft flyovers. The focus of this work is on the calibration of the 3D audio system, including gains used in the amplitude panning algorithm, speaker equalization, and absolute gain control. Because the speakers are installed in an irregularly shaped room, the speaker equalization includes time delay and gain compensation due to different mounting distances from the focal point, filtering for color compensation due to different installations (half space, corner, baffled/unbaffled), and cross-over filtering.
Goverts, S Theo; Huysmans, Elke; Kramer, Sophia E; de Groot, Annette M B; Houtgast, Tammo
2011-12-01
Researchers have used the distortion-sensitivity approach in the psychoacoustical domain to investigate the role of auditory processing abilities in speech perception in noise (van Schijndel, Houtgast, & Festen, 2001; Goverts & Houtgast, 2010). In this study, the authors examined the potential applicability of the distortion-sensitivity approach for investigating the role of linguistic abilities in speech understanding in noise. The authors applied the distortion-sensitivity approach by measuring the processing of visually presented masked text in a condition with manipulated syntactic, lexical, and semantic cues and while using the Text Reception Threshold (George et al., 2007; Kramer, Zekveld, & Houtgast, 2009; Zekveld, George, Kramer, Goverts, & Houtgast, 2007) method. Two groups that differed in linguistic abilities were studied: 13 native and 10 non-native speakers of Dutch, all typically hearing university students. As expected, the non-native subjects showed substantially reduced performance. The results of the distortion-sensitivity approach yielded differentiated results on the use of specific linguistic cues in the 2 groups. The results show the potential value of the distortion-sensitivity approach in studying the role of linguistic abilities in speech understanding in noise of individuals with hearing impairment.
Laboratory Headphone Studies of Human Response to Low-Amplitude Sonic Booms and Rattle Heard Indoors
NASA Technical Reports Server (NTRS)
Loubeau, Alexandra; Sullivan, Brenda M.; Klos, Jacob; Rathsam, Jonathan; Gavin, Joseph R.
2013-01-01
Human response to sonic booms heard indoors is affected by the generation of contact-induced rattle noise. The annoyance caused by sonic boom-induced rattle noise was studied in a series of psychoacoustics tests. Stimuli were divided into three categories and presented in three different studies: isolated rattles at the same calculated Perceived Level (PL), sonic booms combined with rattles with the mixed sound at a single PL, and sonic booms combined with rattles with the mixed sound at three different PL. Subjects listened to sounds over headphones and were asked to report their annoyance. Annoyance to different rattles was shown to vary significantly according to rattle object size. In addition, the combination of low-amplitude sonic booms and rattles can be more annoying than the sonic boom alone. Correlations and regression analyses for the combined sonic boom and rattle sounds identified the Moore and Glasberg Stationary Loudness (MGSL) metric as a primary predictor of annoyance for the tested sounds. Multiple linear regression models were developed to describe annoyance to the tested sounds, and simplifications for applicability to a wider range of sounds are presented.
Jürgens, Tim; Clark, Nicholas R; Lecluyse, Wendy; Meddis, Ray
2016-01-01
To use a computer model of impaired hearing to explore the effects of a physiologically-inspired hearing-aid algorithm on a range of psychoacoustic measures. A computer model of a hypothetical impaired listener's hearing was constructed by adjusting parameters of a computer model of normal hearing. Absolute thresholds, estimates of compression, and frequency selectivity (summarized to a hearing profile) were assessed using this model with and without pre-processing the stimuli by a hearing-aid algorithm. The influence of different settings of the algorithm on the impaired profile was investigated. To validate the model predictions, the effect of the algorithm on hearing profiles of human impaired listeners was measured. A computer model simulating impaired hearing (total absence of basilar membrane compression) was used, and three hearing-impaired listeners participated. The hearing profiles of the model and the listeners showed substantial changes when the test stimuli were pre-processed by the hearing-aid algorithm. These changes consisted of lower absolute thresholds, steeper temporal masking curves, and sharper psychophysical tuning curves. The hearing-aid algorithm affected the impaired hearing profile of the model to approximate a normal hearing profile. Qualitatively similar results were found with the impaired listeners' hearing profiles.
Vibration characteristics of bone conducted sound in vitro.
Stenfelt, S; Håkansson, B; Tjellström, A
2000-01-01
A dry skull added with damping material was used to investigate the vibratory pattern of bone conducted sound. Three orthogonal vibration responses of the cochleae were measured, by means of miniature accelerometers, in the frequency range 0.1-10 kHz. The exciter was attached to the temporal, parietal, and frontal bones, one at the time. In the transmission response to the ipsilateral cochlea, a profound low frequency antiresonance (attenuation) was found, verified psycho-acoustically, and shown to yield a distinct lateralization effect. It was also shown that, for the ipsilateral side, the direction of excitation coincides with that of maximum response. At the contralateral cochlea, no such dominating response direction was found for frequencies above the first skull resonance. An overall higher response level was achieved, for the total energy transmission in general and specifically for the direction of excitation, at the ipsilateral cochlea when the transducer was attached to the excitation point closest to the cochlea. The transranial attenuation was found to be frequency dependent, with values from -5 to 10 dB for the energy transmission and -30 to 40 dB for measurements in a single direction, with a tendency toward higher attenuation at the higher frequencies.
Digital audio watermarking using moment-preserving thresholding
NASA Astrophysics Data System (ADS)
Choi, DooSeop; Jung, Hae Kyung; Choi, Hyuk; Kim, Taejeong
2007-09-01
The Moment-Preserving Thresholding technique for digital images has been used in digital image processing for decades, especially in image binarization and image compression. Its main strength lies in that the binary values that the MPT produces as a result, called representative values, are usually unaffected when the signal being thresholded goes through a signal processing operation. The two representative values in MPT together with the threshold value are obtained by solving the system of the preservation equations for the first, second, and third moment. Relying on this robustness of the representative values to various signal processing attacks considered in the watermarking context, this paper proposes a new watermarking scheme for audio signals. The watermark is embedded in the root-sum-square (RSS) of the two representative values of each signal block using the quantization technique. As a result, the RSS values are modified by scaling the signal according to the watermark bit sequence under the constraint of inaudibility relative to the human psycho-acoustic model. We also address and suggest solutions to the problem of synchronization and power scaling attacks. Experimental results show that the proposed scheme maintains high audio quality and robustness to various attacks including MP3 compression, re-sampling, jittering, and, DA/AD conversion.
USAF Bioenvironmental Noise Data Handbook. Volume 160: KC-10A aircraft, near and far-field noise
NASA Astrophysics Data System (ADS)
Powell, R. G.
1982-09-01
The USAF KC-10A aircraft is an advanced tanker/cargo aircraft powered by three CF6-50C2 turbofan engines. This report provides measured and extrapolated data defining the bioacoustic environments produced by this aircraft operating on a concrete runup pad for eight engine/power configurations. Near-field data are reported for one location in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference levels, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 15 locations are normalized to standard meteorological conditions and extrapolated from 75-8000 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source. Refer to Volume 1 of this handbook, USAF Bioenvironmental Noise Data Handbook, Vol 1: Organization, Content and Application, AMRL-TR-75-50(1) 1975, for discussion of the objective and design of the handbook, the types of data presented, measurement procedures, instrumentation, data processing, definitions of quantities, symbols, equations, applications, limitations, etc.
A multidisciplinary approach of the problem of noise nuisance in urban environment
NASA Astrophysics Data System (ADS)
Rabah, Derbal Cobis; Hamza, Zeghlache
2002-05-01
More often the problem of noise and sonic pollution, particularly in urban sites, is studied by different disciplines such as physics, the acoustics, the psychoacoustics, the medicine and others. It is independently of each other that these sciences are often approaching this subject. Some studies are carried out in laboratories taking noise as samples cut off their realistic context. Urban noise is studied as well by making an abstraction of the different contextual parameters by idealizing a rather complex sonic environment. The noise, according to this present approach, is suposed to react with surounding space, and it takes the form and the quality of the place by defining and requalifying it. It is found that the contextual aspects such as social, cultural or even symbolic dimensions modulate the listening conditions and the perception quality of the noise and even the living and the daily practice of the urban space. The multiparameter dimension study of the noise in an urban context is necessary to better work out the problem and to try to come up with some practical and efficient solutions. The little amount of studies based on such multidisciplinary approach, confort well our effort to go ahead with this methodological approach.
Demonstrations of simple and complex auditory psychophysics for multiple platforms and environments
NASA Astrophysics Data System (ADS)
Horowitz, Seth S.; Simmons, Andrea M.; Blue, China
2005-09-01
Sound is arguably the most widely perceived and pervasive form of energy in our world, and among the least understood, in part due to the complexity of its underlying principles. A series of interactive displays has been developed which demonstrates that the nature of sound involves the propagation of energy through space, and illustrates the definition of psychoacoustics, which is how listeners map the physical aspects of sound and vibration onto their brains. These displays use auditory illusions and commonly experienced music and sound in novel presentations (using interactive computer algorithms) to show that what you hear is not always what you get. The areas covered in these demonstrations range from simple and complex auditory localization, which illustrate why humans are bad at echolocation but excellent at determining the contents of auditory space, to auditory illusions that manipulate fine phase information and make the listener think their head is changing size. Another demonstration shows how auditory and visual localization coincide and sound can be used to change visual tracking. These demonstrations are designed to run on a wide variety of student accessible platforms including web pages, stand-alone presentations, or even hardware-based systems for museum displays.
Varnet, Léo; Knoblauch, Kenneth; Serniclaes, Willy; Meunier, Fanny; Hoen, Michel
2015-01-01
Although there is a large consensus regarding the involvement of specific acoustic cues in speech perception, the precise mechanisms underlying the transformation from continuous acoustical properties into discrete perceptual units remains undetermined. This gap in knowledge is partially due to the lack of a turnkey solution for isolating critical speech cues from natural stimuli. In this paper, we describe a psychoacoustic imaging method known as the Auditory Classification Image technique that allows experimenters to estimate the relative importance of time-frequency regions in categorizing natural speech utterances in noise. Importantly, this technique enables the testing of hypotheses on the listening strategies of participants at the group level. We exemplify this approach by identifying the acoustic cues involved in da/ga categorization with two phonetic contexts, Al- or Ar-. The application of Auditory Classification Images to our group of 16 participants revealed significant critical regions on the second and third formant onsets, as predicted by the literature, as well as an unexpected temporal cue on the first formant. Finally, through a cluster-based nonparametric test, we demonstrate that this method is sufficiently sensitive to detect fine modifications of the classification strategies between different utterances of the same phoneme.
Adductor spasmodic dysphonia: Relationships between acoustic indices and perceptual judgments
NASA Astrophysics Data System (ADS)
Cannito, Michael P.; Sapienza, Christine M.; Woodson, Gayle; Murry, Thomas
2003-04-01
This study investigated relationships between acoustical indices of spasmodic dysphonia and perceptual scaling judgments of voice attributes made by expert listeners. Audio-recordings of The Rainbow Passage were obtained from thirty one speakers with spasmodic dysphonia before and after a BOTOX injection of the vocal folds. Six temporal acoustic measures were obtained across 15 words excerpted from each reading sample, including both frequency of occurrence and percent time for (1) aperiodic phonation, (2) phonation breaks, and (3) fundamental frequency shifts. Visual analog scaling judgments were also obtained from six voice experts using an interactive computer interface to quantify four voice attributes (i.e., overall quality, roughness, brokenness, breathiness) in a carefully psychoacoustically controlled environment, using the same reading passages as stimuli. Number and percent aperiodicity and phonation breaks correlated significanly with perceived overall voice quality, roughness, and brokenness before and after the BOTOX injection. Breathiness was correlated with aperidocity only prior to injection, while roughness also correlated with frequency shifts following injection. Factor analysis reduced perceived attributes to two principal components: glottal squeezing and breathiness. The acoustic measures demonstrated a strong regression relationship with perceived glottal squeezing, but no regression relationship with breathiness was observed. Implications for an analysis of pathologic voices will be discussed.
A consideration on physical tuning for acoustical coloration in recording studio
NASA Astrophysics Data System (ADS)
Shimizu, Yasushi
2003-04-01
Coloration due to particular architectural shapes and dimension or less surface absorption has been mentioned as an acoustical defect in recording studio. Generally interference among early reflected sounds arriving within 10 ms in delay after the direct sound produces coloration by comb filter effect over mid- and high-frequency sounds. In addition, less absorbed room resonance modes also have been well known as a major component for coloration in low-frequency sounds. Small size in dimension with recording studio, however, creates difficulty in characterization associated with wave acoustics behavior, that make acoustical optimization more difficult than that of concert hall acoustics. There still remains difficulty in evaluating amount of coloration as well as predicting its acoustical characteristics in acoustical modeling and in other words acoustical tuning technique during construction is regarded as important to optimize acoustics appropriately to the function of recording studio. This paper presents a example of coloration by comb filtering effect and less damped room modes in typical post-processing recording studio. And acoustical design and measurement technique will be presented for adjusting timbre due to coloration based on psycho-acoustical performance with binaural hearing and room resonance control with line array resonator adjusted to the particular room modes considered.
Richards, Susan; Goswami, Usha
2015-08-01
We investigated whether impaired acoustic processing is a factor in developmental language disorders. The amplitude envelope of the speech signal is known to be important in language processing. We examined whether impaired perception of amplitude envelope rise time is related to impaired perception of lexical and phrasal stress in children with specific language impairment (SLI). Twenty-two children aged between 8 and 12 years participated in this study. Twelve had SLI; 10 were typically developing controls. All children completed psychoacoustic tasks measuring rise time, intensity, frequency, and duration discrimination. They also completed 2 linguistic stress tasks measuring lexical and phrasal stress perception. The SLI group scored significantly below the typically developing controls on both stress perception tasks. Performance on stress tasks correlated with individual differences in auditory sensitivity. Rise time and frequency thresholds accounted for the most unique variance. Digit Span also contributed to task success for the SLI group. The SLI group had difficulties with both acoustic and stress perception tasks. Our data suggest that poor sensitivity to amplitude rise time and sound frequency significantly contributes to the stress perception skills of children with SLI. Other cognitive factors such as phonological memory are also implicated.
Current audiological diagnostics
Hoth, Sebastian; Baljić, Izet
2017-01-01
Today’s audiological functional diagnostics is based on a variety of hearing tests, whose large number takes account of the variety of malfunctions of a complex sensory organ system and the necessity to examine it in a differentiated manner and at any age of life. The objective is to identify nature and origin of the hearing loss and to quantify its extent as far as necessary to dispose of the information needed to initiate the adequate medical (conservative or operational) treatment or the provision with technical hearing aids or prostheses. Moreover, audiometry provides the basis for the assessment of impairment and handicap as well as for the calculation of the degree of disability. In the present overview, the current state of the method inventory available for practical use is described, starting from basic diagnostics over to complex special techniques. The presentation is systematically grouped in subjective procedures, based on psychoacoustic exploration, and objective methods, based on physical measurements: preliminary hearing tests, pure tone threshold, suprathreshold processing of sound intensity, directional hearing, speech understanding in quiet and in noise, dichotic hearing, tympanogram, acoustic reflex, otoacoustic emissions and auditory evoked potentials. Apart from a few still existing gaps, this method inventory covers the whole spectrum of all clinically relevant functional deficits of the auditory system. PMID:29279727
L1 literacy affects L2 pronunciation intake and text vocalization
NASA Astrophysics Data System (ADS)
Walton, Martin
2005-04-01
For both deaf and hearing learners, L1 acquisition calls on auditive, gestural and visual modes in progressive processes over longer stages imposed in strictly anatomical and social order from the earliest pre-lexical phase [Jusczyk (1993), Kuhl & Meltzoff (1996)] to ultimate literacy. By contrast, L2 learning will call on accelerating procedures but with restricted input, arbitrated by L1 literacy as can be traced in the English of French-speaking learners, whether observed in spontaneous speech or in text vocalization modes. An inventory of their predictable omissions, intrusions and substitutions at suprasegmental and syllabic levels, many of which they can actually hear while unable to vocalize in real-time, suggests that a photogenic segmentation of continuous speech into alphabetical units has eclipsed the indispensable earlier phonogenic module, filtering L2 intake and output. This competing mode analysis hypothesizes a critical effect on L2 pronunciation of L1 graphemic procedures acquired usually before puberty, informing data for any Critical Period Hypothesis or amounts of L1 activation influencing L2 accent [Flege (1997, 1998)] or any psychoacoustic French deafness with regard to English stress-timing [Dupoux (1997)]. A metaphonic model [Howell & Dean (1991)] adapted for French learners may remedially distance L1 from L2 vocalization procedures.
Preliminary work about the reproduction of sonic boom signals for perception studies
NASA Astrophysics Data System (ADS)
Epain, N.; Herzog, P.; Rabau, G.; Friot, E.
2006-05-01
As part of a French research program, a sound restitution cabin was designed for investigating the annoyance of sonic boom signals. The first goal was to reproduce the boom spectrum and temporal waveform: this required linear generation of high pressure levels at infrasonic frequencies (110 SPL dB around 3 Hz), and response equalization over the full frequency range (1 Hz-20 kHz). At this stage the pressure inside the cabin was almost uniform around the listener, emulating an outdoor situation. A psychoacoustic study was then conducted which confirmed that the loudness (related to annoyance) of N-waves is roughly governed by the peak pressure, the rise/fall time, and the wave duration. A longer-term goal is to reproduce other aspects of an indoor situation including rattle noise, ground vibrations, and a more realistic spatial repartition of pressure. This latter point has been addressed through an Active Noise Control study aiming at monitoring the low-frequency acoustic pressure on a surface enclosing a listener. Frequency and time-domain numerical simulations of boom reproduction via ANC are given, including a sensitivity study of the coupling between a listener's head and the incident boom wave which combine into the effective sound-field to be reproduced.
Simultaneous masking additivity for short Gaussian-shaped tones: spectral effects.
Laback, Bernhard; Necciari, Thibaud; Balazs, Peter; Savel, Sophie; Ystad, Sølvi
2013-08-01
Laback et al. [(2011). J. Acoust. Soc. Am. 129, 888-897] investigated the additivity of nonsimultaneous masking using short Gaussian-shaped tones as maskers and target. The present study involved Gaussian stimuli to measure the additivity of simultaneous masking for combinations of up to four spectrally separated maskers. According to most basilar membrane measurements, the maskers should be processed linearly at the characteristic frequency (CF) of the target. Assuming also compression of the target, all masker combinations should produce excess masking (exceeding linear additivity). The results for a pair of maskers flanking the target indeed showed excess masking. The amount of excess masking could be predicted by a model assuming summation of masker-evoked excitations in intensity units at the target CF and compression of the target, using compressive input/output functions derived from the nonsimultaneous masking study. However, the combinations of lower-frequency maskers showed much less excess masking than predicted by the model. This cannot easily be attributed to factors like off-frequency listening, combination tone perception, or between-masker suppression. It was better predicted, however, by assuming weighted intensity summation of masker excitations. The optimum weights for the lower-frequency maskers were smaller than one, consistent with partial masker compression as indicated by recent psychoacoustic data.
NASA Astrophysics Data System (ADS)
Kim, Sungyoung; Martens, William L.
2005-04-01
By industry standard (ITU-R. Recommendation BS.775-1), multichannel stereophonic signals within the frequency range of up to 80 or 120 Hz may be mixed and delivered via a single driver (e.g., a subwoofer) without significant impairment of stereophonic sound quality. The assumption that stereophonic information within such low-frequency content is not significant was tested by measuring discrimination thresholds for changes in interaural cross-correlation (IACC) within spectral bands containing the lowest frequency components of low-pitch musical tones. Performances were recorded for three different musical instruments playing single notes ranging in fundamental frequency from 41 Hz to 110 Hz. The recordings, made using a multichannel microphone array composed of five DPA 4006 pressure microphones, were processed to produce a set of stimuli that varied in interaural cross-correlation (IACC) within a low-frequency band, but were otherwise identical in a higher-frequency band. This correlation processing was designed to have minimal effect upon other psychoacoustic variables such as loudness and timbre. The results show that changes in interaural cross correlation (IACC) within low-frequency bands of low-pitch musical tones are most easily discriminated when decorrelated signals are presented via subwoofers positioned at extreme lateral angles (far from the median plane). [Work supported by VRQ.
Derivation of an optimal directivity pattern for sweet spot widening in stereo sound reproduction
NASA Astrophysics Data System (ADS)
Ródenas, Josep A.; Aarts, Ronald M.; Janssen, A. J. E. M.
2003-01-01
In this paper the correction of the degradation of the stereophonic illusion during sound reproduction due to off-center listening is investigated. The main idea is that the directivity pattern of a loudspeaker array should have a well-defined shape such that a good stereo reproduction is achieved in a large listening area. Therefore, a mathematical description to derive an optimal directivity pattern opt that achieves sweet spot widening in a large listening area for stereophonic sound applications is described. This optimal directivity pattern is based on parametrized time/intensity trading data coming from psycho-acoustic experiments within a wide listening area. After the study, the required digital FIR filters are determined by means of a least-squares optimization method for a given stereo base setup (two pair of drivers for the loudspeaker arrays and 2.5-m distance between loudspeakers), which radiate sound in a broad range of listening positions in accordance with the derived opt. Informal listening tests have shown that the opt worked as predicted by the theoretical simulations. They also demonstrated the correct central sound localization for speech and music for a number of listening positions. This application is referred to as ``Position-Independent (PI) stereo.''
A State-of-the-Art Review: Personalization of Tinnitus Sound Therapy.
Searchfield, Grant D; Durai, Mithila; Linford, Tania
2017-01-01
Background: There are several established, and an increasing number of putative, therapies using sound to treat tinnitus. There appear to be few guidelines for sound therapy selection and application. Aim: To review current approaches to personalizing sound therapy for tinnitus. Methods: A "state-of-the-art" review (Grant and Booth, 2009) was undertaken to answer the question: how do current sound-based therapies for tinnitus adjust for tinnitus heterogeneity? Scopus, Google Scholar, Embase and PubMed were searched for the 10-year period 2006-2016. The search strategy used the following key words: "tinnitus" AND "sound" AND "therapy" AND "guidelines" OR "personalized" OR "customized" OR "individual" OR "questionnaire" OR "selection." The results of the review were cataloged and organized into themes. Results: In total 165 articles were reviewed in full, 83 contained sufficient details to contribute to answering the study question. The key themes identified were hearing compensation, pitched-match therapy, maskability, reaction to sound and psychosocial factors. Although many therapies mentioned customization, few could be classified as being personalized. Several psychoacoustic and questionnaire-based methods for assisting treatment selection were identified. Conclusions: Assessment methods are available to assist clinicians to personalize sound-therapy and empower patients to be active in therapy decision-making. Most current therapies are modified using only one characteristic of the individual and/or their tinnitus.
NASA Astrophysics Data System (ADS)
Shinn-Cunningham, Barbara
2003-04-01
One of the key functions of hearing is to help us monitor and orient to events in our environment (including those outside the line of sight). The ability to compute the spatial location of a sound source is also important for detecting, identifying, and understanding the content of a sound source, especially in the presence of competing sources from other positions. Determining the spatial location of a sound source poses difficult computational challenges; however, we perform this complex task with proficiency, even in the presence of noise and reverberation. This tutorial will review the acoustic, psychoacoustic, and physiological processes underlying spatial auditory perception. First, the tutorial will examine how the many different features of the acoustic signals reaching a listener's ears provide cues for source direction and distance, both in anechoic and reverberant space. Then we will discuss psychophysical studies of three-dimensional sound localization in different environments and the basic neural mechanisms by which spatial auditory cues are extracted. Finally, ``virtual reality'' approaches for simulating sounds at different directions and distances under headphones will be reviewed. The tutorial will be structured to appeal to a diverse audience with interests in all fields of acoustics and will incorporate concepts from many areas, such as psychological and physiological acoustics, architectural acoustics, and signal processing.
Shinn-Cunningham, Barbara
2017-10-17
This review provides clinicians with an overview of recent findings relevant to understanding why listeners with normal hearing thresholds (NHTs) sometimes suffer from communication difficulties in noisy settings. The results from neuroscience and psychoacoustics are reviewed. In noisy settings, listeners focus their attention by engaging cortical brain networks to suppress unimportant sounds; they then can analyze and understand an important sound, such as speech, amidst competing sounds. Differences in the efficacy of top-down control of attention can affect communication abilities. In addition, subclinical deficits in sensory fidelity can disrupt the ability to perceptually segregate sound sources, interfering with selective attention, even in listeners with NHTs. Studies of variability in control of attention and in sensory coding fidelity may help to isolate and identify some of the causes of communication disorders in individuals presenting at the clinic with "normal hearing." How well an individual with NHTs can understand speech amidst competing sounds depends not only on the sound being audible but also on the integrity of cortical control networks and the fidelity of the representation of suprathreshold sound. Understanding the root cause of difficulties experienced by listeners with NHTs ultimately can lead to new, targeted interventions that address specific deficits affecting communication in noise. http://cred.pubs.asha.org/article.aspx?articleid=2601617.
Stilp, Christian E.; Kluender, Keith R.
2012-01-01
To the extent that sensorineural systems are efficient, redundancy should be extracted to optimize transmission of information, but perceptual evidence for this has been limited. Stilp and colleagues recently reported efficient coding of robust correlation (r = .97) among complex acoustic attributes (attack/decay, spectral shape) in novel sounds. Discrimination of sounds orthogonal to the correlation was initially inferior but later comparable to that of sounds obeying the correlation. These effects were attenuated for less-correlated stimuli (r = .54) for reasons that are unclear. Here, statistical properties of correlation among acoustic attributes essential for perceptual organization are investigated. Overall, simple strength of the principal correlation is inadequate to predict listener performance. Initial superiority of discrimination for statistically consistent sound pairs was relatively insensitive to decreased physical acoustic/psychoacoustic range of evidence supporting the correlation, and to more frequent presentations of the same orthogonal test pairs. However, increased range supporting an orthogonal dimension has substantial effects upon perceptual organization. Connectionist simulations and Eigenvalues from closed-form calculations of principal components analysis (PCA) reveal that perceptual organization is near-optimally weighted to shared versus unshared covariance in experienced sound distributions. Implications of reduced perceptual dimensionality for speech perception and plausible neural substrates are discussed. PMID:22292057
Evoked-potential changes following discrimination learning involving complex sounds
Orduña, Itzel; Liu, Estella H.; Church, Barbara A.; Eddins, Ann C.; Mercado, Eduardo
2011-01-01
Objective Perceptual sensitivities are malleable via learning, even in adults. We trained adults to discriminate complex sounds (periodic, frequency-modulated sweep trains) using two different training procedures, and used psychoacoustic tests and evoked potential measures (the N1-P2 complex) to assess changes in both perceptual and neural sensitivities. Methods Training took place either on a single day, or daily across eight days, and involved discrimination of pairs of stimuli using a single-interval, forced-choice task. In some participants, training started with dissimilar pairs that became progressively more similar across sessions, whereas in others training was constant, involving only one, highly similar, stimulus pair. Results Participants were better able to discriminate the complex sounds after training, particularly after progressive training, and the evoked potentials elicited by some of the sounds increased in amplitude following training. Significant amplitude changes were restricted to the P2 peak. Conclusion Our findings indicate that changes in perceptual sensitivities parallel enhanced neural processing. Significance These results are consistent with the proposal that changes in perceptual abilities arise from the brain’s capacity to adaptively modify cortical representations of sensory stimuli, and that different training regimens can lead to differences in cortical sensitivities, even after relatively short periods of training. PMID:21958655
Examining explanations for fundamental frequency's contribution to speech intelligibility in noise
NASA Astrophysics Data System (ADS)
Schlauch, Robert S.; Miller, Sharon E.; Watson, Peter J.
2005-09-01
Laures and Weismer [JSLHR, 42, 1148 (1999)] reported that speech with natural variation in fundamental frequency (F0) is more intelligible in noise than speech with a flattened F0 contour. Cognitive-linguistic based explanations have been offered to account for this drop in intelligibility for the flattened condition, but a lower-level mechanism related to auditory streaming may be responsible. Numerous psychoacoustic studies have demonstrated that modulating a tone enables a listener to segregate it from background sounds. To test these rival hypotheses, speech recognition in noise was measured for sentences with six different F0 contours: unmodified, flattened at the mean, natural but exaggerated, reversed, and frequency modulated (rates of 2.5 and 5.0 Hz). The 180 stimulus sentences were produced by five talkers (30 sentences per condition). Speech recognition for fifteen listeners replicate earlier findings showing that flattening the F0 contour results in a roughly 10% reduction in recognition of key words compared with the natural condition. Although the exaggerated condition produced results comparable to those of the flattened condition, the other conditions with unnatural F0 contours all yielded significantly poorer performance than the flattened condition. These results support the cognitive, linguistic-based explanations for the reduction in performance.
Virtual Acoustics: Evaluation of Psychoacoustic Parameters
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Null, Cynthia H. (Technical Monitor)
1997-01-01
Current virtual acoustic displays for teleconferencing and virtual reality are usually limited to very simple or non-existent renderings of reverberation, a fundamental part of the acoustic environmental context that is encountered in day-to-day hearing. Several research efforts have produced results that suggest that environmental cues dramatically improve perceptual performance within virtual acoustic displays, and that is possible to manipulate signal processing parameters to effectively reproduce important aspects of virtual acoustic perception in real-time. However, the computational resources for rendering reverberation remain formidable. Our efforts at NASA Ames have been focused using a several perceptual threshold metrics, to determine how various "trade-offs" might be made in real-time acoustic rendering. This includes both original work and confirmation of existing data that was obtained in real rather than virtual environments. The talk will consider the importance of using individualized versus generalized pinnae cues (the "Head-Related Transfer Function"); the use of head movement cues; threshold data for early reflections and late reverberation; and consideration of the necessary accuracy for measuring and rendering octave-band absorption characteristics of various wall surfaces. In addition, a consideration of the analysis-synthesis of the reverberation within "everyday spaces" (offices, conference rooms) will be contrasted to the commonly used paradigm of concert hall spaces.
Kondaurova, Maria V; Francis, Alexander L
2008-12-01
Two studies explored the role of native language use of an acoustic cue, vowel duration, in both native and non-native contexts in order to test the hypothesis that non-native listeners' reliance on vowel duration instead of vowel quality to distinguish the English tense/lax vowel contrast could be explained by the role of duration as a cue in native phonological contrasts. In the first experiment, native Russian, Spanish, and American English listeners identified stimuli from a beat/bit continuum varying in nine perceptually equal spectral and duration steps. English listeners relied predominantly on spectrum, but showed some reliance on duration. Russian and Spanish speakers relied entirely on duration. In the second experiment, three tests examined listeners' use of vowel duration in native contrasts. Duration was equally important for the perception of lexical stress for all three groups. However, English listeners relied more on duration as a cue to postvocalic consonant voicing than did native Spanish or Russian listeners, and Spanish listeners relied on duration more than did Russian listeners. Results suggest that, although allophonic experience may contribute to cross-language perceptual patterns, other factors such as the application of statistical learning mechanisms and the influence of language-independent psychoacoustic proclivities cannot be ruled out.
Absence of both auditory evoked potentials and auditory percepts dependent on timing cues.
Starr, A; McPherson, D; Patterson, J; Don, M; Luxford, W; Shannon, R; Sininger, Y; Tonakawa, L; Waring, M
1991-06-01
An 11-yr-old girl had an absence of sensory components of auditory evoked potentials (brainstem, middle and long-latency) to click and tone burst stimuli that she could clearly hear. Psychoacoustic tests revealed a marked impairment of those auditory perceptions dependent on temporal cues, that is, lateralization of binaural clicks, change of binaural masked threshold with changes in signal phase, binaural beats, detection of paired monaural clicks, monaural detection of a silent gap in a sound, and monaural threshold elevation for short duration tones. In contrast, auditory functions reflecting intensity or frequency discriminations (difference limens) were only minimally impaired. Pure tone audiometry showed a moderate (50 dB) bilateral hearing loss with a disproportionate severe loss of word intelligibility. Those auditory evoked potentials that were preserved included (1) cochlear microphonics reflecting hair cell activity; (2) cortical sustained potentials reflecting processing of slowly changing signals; and (3) long-latency cognitive components (P300, processing negativity) reflecting endogenous auditory cognitive processes. Both the evoked potential and perceptual deficits are attributed to changes in temporal encoding of acoustic signals perhaps occurring at the synapse between hair cell and eighth nerve dendrites. The results from this patient are discussed in relation to previously published cases with absent auditory evoked potentials and preserved hearing.
Kondaurova, Maria V.; Francis, Alexander L.
2008-01-01
Two studies explored the role of native language use of an acoustic cue, vowel duration, in both native and non-native contexts in order to test the hypothesis that non-native listeners’ reliance on vowel duration instead of vowel quality to distinguish the English tense∕lax vowel contrast could be explained by the role of duration as a cue in native phonological contrasts. In the first experiment, native Russian, Spanish, and American English listeners identified stimuli from a beat∕bit continuum varying in nine perceptually equal spectral and duration steps. English listeners relied predominantly on spectrum, but showed some reliance on duration. Russian and Spanish speakers relied entirely on duration. In the second experiment, three tests examined listeners’ use of vowel duration in native contrasts. Duration was equally important for the perception of lexical stress for all three groups. However, English listeners relied more on duration as a cue to postvocalic consonant voicing than did native Spanish or Russian listeners, and Spanish listeners relied on duration more than did Russian listeners. Results suggest that, although allophonic experience may contribute to cross-language perceptual patterns, other factors such as the application of statistical learning mechanisms and the influence of language-independent psychoacoustic proclivities cannot be ruled out. PMID:19206820
Wendt, Dorothea; Brand, Thomas; Kollmeier, Birger
2014-01-01
An eye-tracking paradigm was developed for use in audiology in order to enable online analysis of the speech comprehension process. This paradigm should be useful in assessing impediments in speech processing. In this paradigm, two scenes, a target picture and a competitor picture, were presented simultaneously with an aurally presented sentence that corresponded to the target picture. At the same time, eye fixations were recorded using an eye-tracking device. The effect of linguistic complexity on language processing time was assessed from eye fixation information by systematically varying linguistic complexity. This was achieved with a sentence corpus containing seven German sentence structures. A novel data analysis method computed the average tendency to fixate the target picture as a function of time during sentence processing. This allowed identification of the point in time at which the participant understood the sentence, referred to as the decision moment. Systematic differences in processing time were observed as a function of linguistic complexity. These differences in processing time may be used to assess the efficiency of cognitive processes involved in resolving linguistic complexity. Thus, the proposed method enables a temporal analysis of the speech comprehension process and has potential applications in speech audiology and psychoacoustics.
Wendt, Dorothea; Brand, Thomas; Kollmeier, Birger
2014-01-01
An eye-tracking paradigm was developed for use in audiology in order to enable online analysis of the speech comprehension process. This paradigm should be useful in assessing impediments in speech processing. In this paradigm, two scenes, a target picture and a competitor picture, were presented simultaneously with an aurally presented sentence that corresponded to the target picture. At the same time, eye fixations were recorded using an eye-tracking device. The effect of linguistic complexity on language processing time was assessed from eye fixation information by systematically varying linguistic complexity. This was achieved with a sentence corpus containing seven German sentence structures. A novel data analysis method computed the average tendency to fixate the target picture as a function of time during sentence processing. This allowed identification of the point in time at which the participant understood the sentence, referred to as the decision moment. Systematic differences in processing time were observed as a function of linguistic complexity. These differences in processing time may be used to assess the efficiency of cognitive processes involved in resolving linguistic complexity. Thus, the proposed method enables a temporal analysis of the speech comprehension process and has potential applications in speech audiology and psychoacoustics. PMID:24950184
Vaerenberg, Bart; Govaerts, Paul J; de Ceulaer, Geert; Daemers, Kristin; Schauwers, Karen
2011-01-01
This report describes the application of the software tool "Fitting to Outcomes eXpert" (FOX) in programming the cochlear implant (CI) processor in new users. FOX is an intelligent agent to assist in the programming of CI processors. The concept of FOX is to modify maps on the basis of specific outcome measures, achieved using heuristic logic and based on a set of deterministic "rules". A prospective study was conducted on eight consecutive CI-users with a follow-up of three months. Eight adult subjects with postlingual deafness were implanted with the Advanced Bionics HiRes90k device. The implants were programmed using FOX, running a set of rules known as Eargroup's EG0910 advice, which features a set of "automaps". The protocol employed for the initial 3 months is presented, with description of the map modifications generated by FOX and the corresponding psychoacoustic test results. The 3 month median results show 25 dBHL as PTA, 77% (55 dBSPL) and 71% (70 dBSPL) phoneme score at speech audiometry and loudness scaling in or near to the normal zone at different frequencies. It is concluded that this approach is feasible to start up CI fitting and yields good outcome.
Bilateral transverse sinus stenosis in patients with tinnitus.
Chiarella, G; Bono, F; Cassandro, C; Lopolito, M; Quattrone, A; Cassandro, E
2012-08-01
Tinnitus is a frequent complaint in patients affected by intracranial hypertension (IH). Recently, some studies have reported an association between idiopathic intracranial hypertension (IIH) and bilateral transverse sinus stenosis (BTSS). We investigated the relationship between BTSS and monosymptomatic tinnitus, regardless of its clinical characteristics, in subjects without clinical evidence of IH. We selected 78 subjects (all women, mean age 49.5 ± 10.36) affected by tinnitus, without clinical history of audiological and otological pathologies, enrolled among outpatients of the Institute of Audiology and Phoniatrics in Catanzaro, Italy, over a 2 year period. All subjects underwent psychometric evaluation, psychoacoustic assessment, neurological and ophthalmological examination, cerebral magnetic resonance venography (MRV) and brain magnetic resonance imaging (MRI). MRV identified BTSS in 17.9% (14 patients). In the BTSS group, tinnitus was bilateral/central in 21%, and monolateral in the remaining patients (50% left; 29% right ear). It was more frequently pulsating in the BTSS group, but 64.9% of BTSS subjects described their tinnitus as stable. No features of tinnitus showed statistical significance in association with BTSS. In BTSS subjects, we found values suggesting IH by lumbar puncture (LP) in 40% of cases. In these patients, LP gave immediate improvement of tinnitus. The association between BTSS and tinnitus, regardless of its features, must be considered when other causes of tinnitus are excluded.
The effects of tones in noise on human annoyance and performance
NASA Astrophysics Data System (ADS)
Lee, Joonhee
Building mechanical equipment often generates prominent tones because most systems include rotating parts like fans and pumps. These tonal noises can cause unpleasant user experiences in spaces and, in turn, lead to increased complaints by building occupants. Currently, architectural engineers can apply the noise criteria guidelines in standards or publications to achieve acceptable noise conditions for assorted types of spaces. However, these criteria do not apply well if the noise contains perceptible tones. The annoyance thresholds experienced by the general population with regards to the degree of tones in noise is a significant piece of knowledge that has not been well-established. Thus, this dissertation addresses the relationship between human perception and noises with tones in the built environment. Four phases of subjective testing were conducted in an indoor acoustic testing chamber at the University of Nebraska to achieve the research objective. The results indicate that even the least prominent tones in noises can significantly decrease the cognitive performance of participants on a mentally demanding task. Factorial repeated-measures analysis of variance of test results have proven that tonality has a crucial influence on working memory capacity of subjects, whereas loudness levels alone did not. A multidimensional annoyance model, incorporating psycho-acoustical attributes of noise in addition to loudness and tonality, has been proposed as a more accurate annoyance model.
NASA Astrophysics Data System (ADS)
Beauchamp, James W.
2002-11-01
Software has been developed which enables users to perform time-varying spectral analysis of individual musical tones or successions of them and to perform further processing of the data. The package, called sndan, is freely available in source code, uses EPS graphics for display, and is written in ansi c for ease of code modification and extension. Two analyzers, a fixed-filter-bank phase vocoder (''pvan'') and a frequency-tracking analyzer (''mqan'') constitute the analysis front end of the package. While pvan's output consists of continuous amplitudes and frequencies of harmonics, mqan produces disjoint ''tracks.'' However, another program extracts a fundamental frequency and separates harmonics from the tracks, resulting in a continuous harmonic output. ''monan'' is a program used to display harmonic data in a variety of formats, perform various spectral modifications, and perform additive resynthesis of the harmonic partials, including possible pitch-shifting and time-scaling. Sounds can also be synthesized according to a musical score using a companion synthesis language, Music 4C. Several other programs in the sndan suite can be used for specialized tasks, such as signal display and editing. Applications of the software include producing specialized sounds for music compositions or psychoacoustic experiments or as a basis for developing new synthesis algorithms.
Cortical encoding of pitch: Recent results and open questions
Walker, Kerry M.M.; Bizley, Jennifer K.; King, Andrew J.; Schnupp, Jan W.H.
2011-01-01
It is widely appreciated that the key predictor of the pitch of a sound is its periodicity. Neural structures which support pitch perception must therefore be able to reflect the repetition rate of a sound, but this alone is not sufficient. Since pitch is a psychoacoustic property, a putative cortical code for pitch must also be able to account for the relationship between the amount to which a sound is periodic (i.e. its temporal regularity) and the perceived pitch salience, as well as limits in our ability to detect pitch changes or to discriminate rising from falling pitch. Pitch codes must also be robust in the presence of nuisance variables such as loudness or timbre. Here, we review a large body of work on the cortical basis of pitch perception, which illustrates that the distribution of cortical processes that give rise to pitch perception is likely to depend on both the acoustical features and functional relevance of a sound. While previous studies have greatly advanced our understanding, we highlight several open questions regarding the neural basis of pitch perception. These questions can begin to be addressed through a cooperation of investigative efforts across species and experimental techniques, and, critically, by examining the responses of single neurons in behaving animals. PMID:20457240
[Interdisciplinary management of chronic tinnitus (II)].
Rosanowski, F; Hoppe, U; Köllner, V; Weber, A; Eysholdt, U
2001-06-01
Tinnitus is defined as the perception of sound in the absence of any appropriate external stimulation. It is a common, bothersome and hard-to-evaluate symptom and in most cases it cannot be objectified. Its incidence in Germany and the western world is about 10%. About 1-2% of the population are severely disturbed by tinnitus and it may disrupt everyday activities and sleep. Recent theoretical developments favour a neurophysiological approach as an explanation for tinnitus in addition to a psychoacoustic model based on peripheral lesion in the cochlea or auditory nerve. In the neurophysiological model, the processing of the tinnitus signal plays a dominant role in its detection, perception and evaluation. Therefore, attention and other psychological factors become important in the understanding and treatment of tinnitus. Many treatments of chronic tinnitus have been proposed and implemented. Today, cognitive-behavioural treatment is regarded as an important part of an integrative therapy which may be compiled of counselling, relaxation therapy, instrumental (hearing aid, tinnitus masker, tinnitus instrument, tinnitus noiser) and pharmacological tools (lidocaine, neurotransmitters). In well-controlled studies the empirical support for other therapeutical approaches such as acupuncture is weak. This work gives a review of the current knowledge of the etiology, pathogenesis, epidemiology, the interdisciplinary diagnostic approach and treatment of tinnitus and especially focuses on insurance and medico-legal aspects.
Comparison of auditory stream segregation in sighted and early blind individuals.
Boroujeni, Fatemeh Moghadasi; Heidari, Fatemeh; Rouzbahani, Masoumeh; Kamali, Mohammad
2017-01-18
An important characteristic of the auditory system is the capacity to analyze complex sounds and make decisions on the source of the constituent parts of these sounds. Blind individuals compensate for the lack of visual information by an increase input from other sensory modalities, including increased auditory information. The purpose of the current study was to compare the fission boundary (FB) threshold of sighted and early blind individuals through spectral aspects using a psychoacoustic auditory stream segregation (ASS) test. This study was conducted on 16 sighted and 16 early blind adult individuals. The applied stimuli were presented sequentially as the pure tones A and B and as a triplet ABA-ABA pattern at the intensity of 40dBSL. The A tone frequency was selected as the basis at values of 500, 1000, and 2000Hz. The B tone was presented with the difference of a 4-100% above the basis tone frequency. Blind individuals had significantly lower FB thresholds than sighted people. FB was independent of the frequency of the tone A when expressed as the difference in the number of equivalent rectangular bandwidths (ERBs). Early blindness may increase perceptual separation of the acoustic stimuli to form accurate representations of the world. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Neighborhood noise pollution as a determinant of displaced aggression: a pilot study.
Dzhambov, Angel; Dimitrova, Donka
2014-01-01
Noise pollution is still a growing public health problem with a significant impact on psychological health and well-being. The aim of this study was to investigate the impact of noise on displaced aggression (DA) in different subgroups of residents in one of the neighborhoods of Plovdiv city. A cross-sectional semi-structured interview survey was conducted using specially designed data registration forms and 33 close-ended and open-ended questions, divided into two major panels - one original and a modified version of the Displaced Aggression Questionnaire (DAQ). The mean score for DA was 61.12 (±19.97). Hearing noises above the perceived normal threshold, higher noise sensitivity and continuous noises were associated with higher levels of DA. Low frequency and high intensity noises were also associated with higher DA scores. Multiple regression model supported these findings. Contradictory to previous research age was positively correlated with noise sensitivity and aggression. We speculated that this might be due to the relatively lower socio-economic standard and quality of life in Bulgaria. Therefore, social climate might be modifying the way people perceive and react to environmental noise. Finally, the DAQ proved to be a viable measurement tool of these associations and might be further implemented and modified to suit the purposes of psychoacoustic assessment.
Acoustic resonance at the dawn of life: musical fundamentals of the psychoanalytic relationship.
Pickering, Judith
2015-11-01
This paper uses a case vignette to show how musical elements of speech are a crucial source of information regarding the patient's emotional states and associated memory systems that are activated at a given moment in the analytic field. There are specific psychoacoustic markers associated with different memory systems which indicate whether a patient is immersed in a state of creative intersubjective relatedness related to autobiographical memory, or has been triggered into a traumatic memory system. When a patient feels immersed in an atmosphere of intersubjective mutuality, dialogue features a rhythmical and tuneful form of speech featuring improvized reciprocal imitation, theme and variation. When the patient is catapulted into a traumatic memory system, speech becomes monotone and disjointed. Awareness of such acoustic features of the traumatic memory system helps to alert the analyst that such a shift has taken place informing appropriate responses and interventions. Communicative musicality (Malloch & Trevarthen 2009) originates in the earliest non-verbal vocal communication between infant and care-giver, states of primary intersubjectivity. Such musicality continues to be the primary vehicle for transmitting emotional meaning and for integrating right and left hemispheres. This enables communication that expresses emotional significance, personal value as well as conceptual reasoning. © 2015, The Society of Analytical Psychology.
Effect of train type on annoyance and acoustic features of the rolling noise.
Kasess, Christian H; Noll, Anton; Majdak, Piotr; Waubke, Holger
2013-08-01
This study investigated the annoyance associated with the rolling noise of different railway stock. Passbys of nine train types (passenger and freight trains) equipped with different braking systems were recorded. Acoustic features showed a clear distinction of the braking system with the A-weighted energy equivalent sound level (LAeq) showing a difference in the range of 10 dB between cast-iron braked trains and trains with disk or K-block brakes. Further, annoyance was evaluated in a psychoacoustic experiment where listeners rated the relative annoyance of the rolling noise for the different train types. Stimuli with and without the original LAeq differences were tested. For the original LAeq differences, the braking system significantly affected the annoyance with cast-iron brakes being most annoying, most likely as a consequence of the increased wheel roughness causing an increased LAeq. Contribution of the acoustic features to the annoyance was investigated revealing that the LAeq explained up to 94% of the variance. For the stimuli without differences in the LAeq, cast-iron braked train types were significantly less annoying and the spectral features explained up to 60% of the variance in the annoyance. The effect of these spectral features on the annoyance of the rolling noise is discussed.
Geravanchizadeh, Masoud; Fallah, Ali
2015-12-01
A binaural and psychoacoustically motivated intelligibility model, based on a well-known monaural microscopic model is proposed. This model simulates a phoneme recognition task in the presence of spatially distributed speech-shaped noise in anechoic scenarios. In the proposed model, binaural advantage effects are considered by generating a feature vector for a dynamic-time-warping speech recognizer. This vector consists of three subvectors incorporating two monaural subvectors to model the better-ear hearing, and a binaural subvector to simulate the binaural unmasking effect. The binaural unit of the model is based on equalization-cancellation theory. This model operates blindly, which means separate recordings of speech and noise are not required for the predictions. Speech intelligibility tests were conducted with 12 normal hearing listeners by collecting speech reception thresholds (SRTs) in the presence of single and multiple sources of speech-shaped noise. The comparison of the model predictions with the measured binaural SRTs, and with the predictions of a macroscopic binaural model called extended equalization-cancellation, shows that this approach predicts the intelligibility in anechoic scenarios with good precision. The square of the correlation coefficient (r(2)) and the mean-absolute error between the model predictions and the measurements are 0.98 and 0.62 dB, respectively.
A neurally inspired musical instrument classification system based upon the sound onset.
Newton, Michael J; Smith, Leslie S
2012-06-01
Physiological evidence suggests that sound onset detection in the auditory system may be performed by specialized neurons as early as the cochlear nucleus. Psychoacoustic evidence shows that the sound onset can be important for the recognition of musical sounds. Here the sound onset is used in isolation to form tone descriptors for a musical instrument classification task. The task involves 2085 isolated musical tones from the McGill dataset across five instrument categories. A neurally inspired tone descriptor is created using a model of the auditory system's response to sound onset. A gammatone filterbank and spiking onset detectors, built from dynamic synapses and leaky integrate-and-fire neurons, create parallel spike trains that emphasize the sound onset. These are coded as a descriptor called the onset fingerprint. Classification uses a time-domain neural network, the echo state network. Reference strategies, based upon mel-frequency cepstral coefficients, evaluated either over the whole tone or only during the sound onset, provide context to the method. Classification success rates for the neurally-inspired method are around 75%. The cepstral methods perform between 73% and 76%. Further testing with tones from the Iowa MIS collection shows that the neurally inspired method is considerably more robust when tested with data from an unrelated dataset.
Whiteford, Kelly L.; Oxenham, Andrew J.
2015-01-01
The question of how frequency is coded in the peripheral auditory system remains unresolved. Previous research has suggested that slow rates of frequency modulation (FM) of a low carrier frequency may be coded via phase-locked temporal information in the auditory nerve, whereas FM at higher rates and/or high carrier frequencies may be coded via a rate-place (tonotopic) code. This hypothesis was tested in a cohort of 100 young normal-hearing listeners by comparing individual sensitivity to slow-rate (1-Hz) and fast-rate (20-Hz) FM at a carrier frequency of 500 Hz with independent measures of phase-locking (using dynamic interaural time difference, ITD, discrimination), level coding (using amplitude modulation, AM, detection), and frequency selectivity (using forward-masking patterns). All FM and AM thresholds were highly correlated with each other. However, no evidence was obtained for stronger correlations between measures thought to reflect phase-locking (e.g., slow-rate FM and ITD sensitivity), or between measures thought to reflect tonotopic coding (fast-rate FM and forward-masking patterns). The results suggest that either psychoacoustic performance in young normal-hearing listeners is not limited by peripheral coding, or that similar peripheral mechanisms limit both high- and low-rate FM coding. PMID:26627783
Spectro-temporal modulation masking patterns reveal frequency selectivity.
Oetjen, Arne; Verhey, Jesko L
2015-02-01
The present study investigated the possibility that the human auditory system demonstrates frequency selectivity to spectro-temporal amplitude modulations. Threshold modulation depth for detecting sinusoidal spectro-temporal modulations was measured using a generalized masked threshold pattern paradigm with narrowband masker modulations. Four target spectro-temporal modulations were examined, differing in their temporal and spectral modulation frequencies: a temporal modulation of -8, 8, or 16 Hz combined with a spectral modulation of 1 cycle/octave and a temporal modulation of 4 Hz combined with a spectral modulation of 0.5 cycles/octave. The temporal center frequencies of the masker modulation ranged from 0.25 to 4 times the target temporal modulation. The spectral masker-modulation center-frequencies were 0, 0.5, 1, 1.5, and 2 times the target spectral modulation. For all target modulations, the pattern of average thresholds for the eight normal-hearing listeners was consistent with the hypothesis of a spectro-temporal modulation filter. Such a pattern of modulation-frequency sensitivity was predicted on the basis of psychoacoustical data for purely temporal amplitude modulations and purely spectral amplitude modulations. An analysis of separability indicates that, for the present data set, selectivity in the spectro-temporal modulation domain can be described by a combination of a purely spectral and a purely temporal modulation filter function.
Whiteford, Kelly L; Oxenham, Andrew J
2015-11-01
The question of how frequency is coded in the peripheral auditory system remains unresolved. Previous research has suggested that slow rates of frequency modulation (FM) of a low carrier frequency may be coded via phase-locked temporal information in the auditory nerve, whereas FM at higher rates and/or high carrier frequencies may be coded via a rate-place (tonotopic) code. This hypothesis was tested in a cohort of 100 young normal-hearing listeners by comparing individual sensitivity to slow-rate (1-Hz) and fast-rate (20-Hz) FM at a carrier frequency of 500 Hz with independent measures of phase-locking (using dynamic interaural time difference, ITD, discrimination), level coding (using amplitude modulation, AM, detection), and frequency selectivity (using forward-masking patterns). All FM and AM thresholds were highly correlated with each other. However, no evidence was obtained for stronger correlations between measures thought to reflect phase-locking (e.g., slow-rate FM and ITD sensitivity), or between measures thought to reflect tonotopic coding (fast-rate FM and forward-masking patterns). The results suggest that either psychoacoustic performance in young normal-hearing listeners is not limited by peripheral coding, or that similar peripheral mechanisms limit both high- and low-rate FM coding.
Toward Reduced Aircraft Community Noise Impact Via a Perception-Influenced Design Approach
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.
2016-01-01
This is an exciting time for aircraft design. New configurations, including small multi-rotor uncrewed aerial systems, fixed- and tilt-wing distributed electric propulsion aircraft, high-speed rotorcraft, hybrid-electric commercial transports, and low-boom supersonic transports, are being made possible through a host of propulsion and airframe technology developments. The resulting noise signatures may be radically different, both spectrally and temporally, than those of the current fleet. Noise certification metrics currently used in aircraft design do not necessarily reflect these characteristics and therefore may not correlate well with human response. Further, as operations and missions become less airport-centric, e.g., those associated with on-demand mobility or package delivery, vehicles may operate in closer proximity to the population than ever before. Fortunately, a new set of tools are available for assessing human perception during the design process in order to affect the final design in a positive manner. The tool chain utilizes system noise prediction methods coupled with auralization and psychoacoustic testing, making possible the inclusion of human response to noise, along with performance criteria and certification requirements, into the aircraft design process. Several case studies are considered to illustrate how this approach could be used to influence the design of future aircraft.
Newman, C W; Wharton, J A; Shivapuja, B G; Jacobson, G P
1994-01-01
Tinnitus is often a disturbing symptom which affects 6-20% of the population. Relationships among tinnitus pitch and loudness judgments, audiometric speech understanding measures and self-perceived handicap were evaluated in a sample of subjects with tinnitus and hearing loss (THL). Data obtained from the THL sample on the audiometric speech measures were compared to the performance of an age-matched hearing loss only (HL) group. Both groups had normal hearing through 1 kHz with a sloping configuration of < or = 20 dB/octave between 2-12 kHz. The THL subjects performed more poorly on the low predictability items of the Speech Perception in Noise Test, suggesting that tinnitus may interfere with the perception of speech signals having reduced linguistic redundancy. The THL subjects rated their tinnitus as annoying at relatively low sensation levels using the pitch-match frequency as the reference tone. Further, significant relationships were found between loudness judgment measures and self-rated annoyance. No predictable relationships were observed between the audiometric speech measures and perceived handicap using the Tinnitus Handicap Questionnaire. These findings support the use of self-report measures in tinnitus patients in that audiometric speech tests alone may be insufficient in describing an individual's reaction to his/her communication breakdowns.
NASA Astrophysics Data System (ADS)
Mohlman, H. T.
1983-04-01
The Air Force community noise prediction model (NOISEMAP) is used to describe the aircraft noise exposure around airbases and thereby aid airbase planners to minimize exposure and prevent community encroachment which could limit mission effectiveness of the installation. This report documents two computer programs (OMEGA 10 and OMEGA 11) which were developed to prepare aircraft flight and ground runup noise data for input to NOISEMAP. OMEGA 10 is for flight operations and OMEGA 11 is for aircraft ground runups. All routines in each program are documented at a level useful to a programmer working with the code or a reader interested in a general overview of what happens within a specific subroutine. Both programs input normalized, reference aircraft noise data; i.e., data at a standard reference distance from the aircraft, for several fixed engine power settings, a reference airspeed and standard day meteorological conditions. Both programs operate on these normalized, reference data in accordance with user-defined, non-reference conditions to derive single-event noise data for 22 distances (200 to 25,000 feet) in a variety of physical and psycho-acoustic metrics. These outputs are in formats ready for input to NOISEMAP.
Derivation of an optimal directivity pattern for sweet spot widening in stereo sound reproduction.
Ródenas, Josep A; Aarts, Ronald M; Janssen, A J E M
2003-01-01
In this paper the correction of the degradation of the stereophonic illusion during sound reproduction due to off-center listening is investigated. The main idea is that the directivity pattern of a loudspeaker array should have a well-defined shape such that a good stereo reproduction is achieved in a large listening area. Therefore, a mathematical description to derive an optimal directivity pattern l(opt) that achieves sweet spot widening in a large listening area for stereophonic sound applications is described. This optimal directivity pattern is based on parametrized time/intensity trading data coming from psycho-acoustic experiments within a wide listening area. After the study, the required digital FIR filters are determined by means of a least-squares optimization method for a given stereo base setup (two pair of drivers for the loudspeaker arrays and 2.5-m distance between loudspeakers), which radiate sound in a broad range of listening positions in accordance with the derived l(opt). Informal listening tests have shown that the l(opt) worked as predicted by the theoretical simulations. They also demonstrated the correct central sound localization for speech and music for a number of listening positions. This application is referred to as "Position-Independent (PI) stereo."
Effect of musical training on static and dynamic measures of spectral-pattern discrimination.
Sheft, Stanley; Smayda, Kirsten; Shafiro, Valeriy; Maddox, W Todd; Chandrasekaran, Bharath
2013-06-01
Both behavioral and physiological studies have demonstrated enhanced processing of speech in challenging listening environments attributable to musical training. The relationship, however, of this benefit to auditory abilities as assessed by psychoacoustic measures remains unclear. Using tasks previously shown to relate to speech-in-noise perception, the present study evaluated discrimination ability for static and dynamic spectral patterns by 49 listeners grouped as either musicians or nonmusicians. The two static conditions measured the ability to detect a change in the phase of a logarithmic sinusoidal spectral ripple of wideband noise with ripple densities of 1.5 and 3.0 cycles per octave chosen to emphasize either timbre or pitch distinctions, respectively. The dynamic conditions assessed temporal-pattern discrimination of 1-kHz pure tones frequency modulated by different lowpass noise samples with thresholds estimated in terms of either stimulus duration or signal-to-noise ratio. Musicians performed significantly better than nonmusicians on all four tasks. Discriminant analysis showed that group membership was correctly predicted for 88% of the listeners with the structure coefficient of each measure greater than 0.51. Results suggest that enhanced processing of static and dynamic spectral patterns defined by low-rate modulation may contribute to the relationship between musical training and speech-in-noise perception. [Supported by NIH.].
Comparison of Interaural Electrode Pairing Methods for Bilateral Cochlear Implants
Dietz, Mathias
2015-01-01
In patients with bilateral cochlear implants (CIs), pairing matched interaural electrodes and stimulating them with the same frequency band is expected to facilitate binaural functions such as binaural fusion, localization, and spatial release from masking. Because clinical procedures typically do not include patient-specific interaural electrode pairing, it remains the case that each electrode is allocated to a generic frequency range, based simply on the electrode number. Two psychoacoustic techniques for determining interaurally paired electrodes have been demonstrated in several studies: interaural pitch comparison and interaural time difference (ITD) sensitivity. However, these two methods are rarely, if ever, compared directly. A third, more objective method is to assess the amplitude of the binaural interaction component (BIC) derived from electrically evoked auditory brainstem responses for different electrode pairings; a method has been demonstrated to be a potential candidate for bilateral CI users. Here, we tested all three measures in the same eight CI users. We found good correspondence between the electrode pair producing the largest BIC and the electrode pair producing the maximum ITD sensitivity. The correspondence between the pairs producing the largest BIC and the pitch-matched electrode pairs was considerably weaker, supporting the previously proposed hypothesis that whilst place pitch might adapt over time to accommodate mismatched inputs, sensitivity to ITDs does not adapt to the same degree. PMID:26631108
Colletti, V; Fiorino, F G
1993-11-01
To facilitate identification and preservation of the auditory nerve during cerebello-pontine angle surgery, bipolar recording of cochlear nerve compound action potentials (CNAPs) was performed. Two silver wires insulated with teflon up to the exposed ends were utilized as electrodes. They were twisted together, the distance between the two tips being 1 mm or less. Rarefaction polarity clicks (31/s) ranging from the psychoacoustical threshold to 120 dB pe SPL were used as stimuli. The investigation was performed in three groups of patients. The first group consisted of 9 patients submitted to vestibular neurectomy and 4 patients operated on by microvascular decompression of the eighth nerve. The second group comprised 8 patients with acoustic tumors smaller than 2 mm and serviceable hearing. Postoperative audiometric results in the subjects in the second group were compared with those obtained in well-matched homogeneous controls consisting of patients with acoustic neuroma operated on without the aid of CNAP recording. Bipolar recording from the eighth nerve was extremely selective, a good response being obtained only when positioning the electrode on the cochlear portion of the eighth nerve. During removal of the acoustic neuroma, repeated bipolar probing of the tumor and eighth nerve facilitated the task of distinguishing the cochlear nerve from other nervous structures and from the tumor, and contributed to preserving hearing in most patients.
2017-01-01
Purpose This review provides clinicians with an overview of recent findings relevant to understanding why listeners with normal hearing thresholds (NHTs) sometimes suffer from communication difficulties in noisy settings. Method The results from neuroscience and psychoacoustics are reviewed. Results In noisy settings, listeners focus their attention by engaging cortical brain networks to suppress unimportant sounds; they then can analyze and understand an important sound, such as speech, amidst competing sounds. Differences in the efficacy of top-down control of attention can affect communication abilities. In addition, subclinical deficits in sensory fidelity can disrupt the ability to perceptually segregate sound sources, interfering with selective attention, even in listeners with NHTs. Studies of variability in control of attention and in sensory coding fidelity may help to isolate and identify some of the causes of communication disorders in individuals presenting at the clinic with “normal hearing.” Conclusions How well an individual with NHTs can understand speech amidst competing sounds depends not only on the sound being audible but also on the integrity of cortical control networks and the fidelity of the representation of suprathreshold sound. Understanding the root cause of difficulties experienced by listeners with NHTs ultimately can lead to new, targeted interventions that address specific deficits affecting communication in noise. Presentation Video http://cred.pubs.asha.org/article.aspx?articleid=2601617 PMID:29049598
Li, Feipeng; Trevino, Andrea; Menon, Anjali; Allen, Jont B
2012-10-01
In a previous study on plosives, the 3-Dimensional Deep Search (3DDS) method for the exploration of the necessary and sufficient cues for speech perception was introduced (Li et al., (2010). J. Acoust. Soc. Am. 127(4), 2599-2610). Here, this method is used to isolate the spectral cue regions for perception of the American English fricatives /∫, 3, s, z, f, v, θ, δ in time, frequency, and intensity. The fricatives are analyzed in the context of consonant-vowel utterances, using the vowel /α/. The necessary cues were found to be contained in the frication noise for /∫, 3, s, z, f, v/. 3DDS analysis isolated the cue regions of /s, z/ between 3.6 and 8 [kHz] and /∫, 3/ between 1.4 and 4.2 [kHz]. Some utterances were found to contain acoustic components that were unnecessary for correct perception, but caused listeners to hear non-target consonants when the primary cue region was removed; such acoustic components are labeled "conflicting cue regions." The amplitude modulation of the high-frequency frication region by the fundamental F0 was found to be a sufficient cue for voicing. Overall, the 3DDS method allows one to analyze the effects of natural speech components without initial assumptions about where perceptual cues lie in time-frequency space or which elements of production they correspond to.
MUSIC APPRECIATION AND TRAINING FOR COCHLEAR IMPLANT RECIPIENTS: A REVIEW
Looi, Valerie; Gfeller, Kate; Driscoll, Virginia
2012-01-01
In recent years, there has been increasing interest in music perception of cochlear implant (CI) recipients, and a growing body of research conducted in this area. The majority of these studies have examined perceptual accuracy for pitch, rhythm, and timbre. Another important, but less commonly studied aspect of music listening is appreciation, or appraisal. Despite the ongoing research into potential technological improvements that may improve music perception for recipients, both perceptual accuracy and appreciation generally remain poor for most recipients. Whilst perceptual accuracy for music is important, appreciation and enjoyment also warrants research as it also contributes to clinical outcomes and perceived benefits. Music training is being shown to offer excellent potential for improving music perception and appreciation for recipients. Therefore, the primary topics of this review are music appreciation and training. However, a brief overview of the psychoacoustic, technical, and physiological factors associated with a recipient’s perception of music is provided, as these are important factors in understanding the listening experience for CI recipients. The purpose of this review is to summarize key papers that have investigated these issues, in order to demonstrate that i) music enjoyment and appraisal is an important and valid consideration in evaluating music outcomes for recipients, and ii) that music training can improve music listening for many recipients, and is something that can be offered to persons using current technology. PMID:23459244
A State-of-the-Art Review: Personalization of Tinnitus Sound Therapy
Searchfield, Grant D.; Durai, Mithila; Linford, Tania
2017-01-01
Background: There are several established, and an increasing number of putative, therapies using sound to treat tinnitus. There appear to be few guidelines for sound therapy selection and application. Aim: To review current approaches to personalizing sound therapy for tinnitus. Methods: A “state-of-the-art” review (Grant and Booth, 2009) was undertaken to answer the question: how do current sound-based therapies for tinnitus adjust for tinnitus heterogeneity? Scopus, Google Scholar, Embase and PubMed were searched for the 10-year period 2006–2016. The search strategy used the following key words: “tinnitus” AND “sound” AND “therapy” AND “guidelines” OR “personalized” OR “customized” OR “individual” OR “questionnaire” OR “selection.” The results of the review were cataloged and organized into themes. Results: In total 165 articles were reviewed in full, 83 contained sufficient details to contribute to answering the study question. The key themes identified were hearing compensation, pitched-match therapy, maskability, reaction to sound and psychosocial factors. Although many therapies mentioned customization, few could be classified as being personalized. Several psychoacoustic and questionnaire-based methods for assisting treatment selection were identified. Conclusions: Assessment methods are available to assist clinicians to personalize sound-therapy and empower patients to be active in therapy decision-making. Most current therapies are modified using only one characteristic of the individual and/or their tinnitus. PMID:28970812
Design guidelines for the use of audio cues in computer interfaces
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sumikawa, D.A.; Blattner, M.M.; Joy, K.I.
1985-07-01
A logical next step in the evolution of the computer-user interface is the incorporation of sound thereby using our senses of ''hearing'' in our communication with the computer. This allows our visual and auditory capacities to work in unison leading to a more effective and efficient interpretation of information received from the computer than by sight alone. In this paper we examine earcons, which are audio cues, used in the computer-user interface to provide information and feedback to the user about computer entities (these include messages and functions, as well as states and labels). The material in this paper ismore » part of a larger study that recommends guidelines for the design and use of audio cues in the computer-user interface. The complete work examines the disciplines of music, psychology, communication theory, advertising, and psychoacoustics to discover how sound is utilized and analyzed in those areas. The resulting information is organized according to the theory of semiotics, the theory of signs, into the syntax, semantics, and pragmatics of communication by sound. Here we present design guidelines for the syntax of earcons. Earcons are constructed from motives, short sequences of notes with a specific rhythm and pitch, embellished by timbre, dynamics, and register. Compound earcons and family earcons are introduced. These are related motives that serve to identify a family of related cues. Examples of earcons are given.« less
Analysis of the relationship between cognitive skills and unilateral sensory hearing loss.
Calderón-Leyva, I; Díaz-Leines, S; Arch-Tirado, E; Lino-González, A L
2018-06-01
To analyse cognitive skills in patients with severe unilateral hearing loss versus those in subjects with normal hearing. 40 adults participated: 20 patients (10 women and 10 men) with severe unilateral hearing loss and 20 healthy subjects matched to the study group. Cognitive abilities were measured with the Spanish version of the Woodcock Johnson Battery-Revised; central auditory processing was assessed with monaural psychoacoustic tests. Box plots were drawn and t tests were performed for samples with a significance of P≤.05. A comparison of performances on the filtered word testing and time-compressed disyllabic word tests between patients and controls revealed a statistically significant difference (P≤.05) with greater variability among responses by hearing impaired subjects. This same group also showed a better cognitive performance on the numbers reversed, visual auditory learning, analysis synthesis, concept formation, and incomplete words tests. Patients with hearing loss performed more poorly than controls on the filtered word and time-compressed disyllabic word tests, but more competently on memory, reasoning, and auditory processing tasks. Complementary tests, such as those assessing central auditory processes and cognitive ability tests, are important and helpful for designing habilitation/rehabilitation and therapeutic strategies intended to optimise and stimulate cognitive skills in subjects with unilateral hearing impairment. Copyright © 2016 Sociedad Española de Neurología. Publicado por Elsevier España, S.L.U. All rights reserved.
Detecting components of hearing aid fitting using a self-assessment-inventory.
Meister, Hartmut; Lausberg, Isabel; Kiessling, Juergen; von Wedel, Hasso; Walger, Martin
2005-07-01
The evaluation of hearing-aid fitting includes numerous assessments such as electro- and psychoacoustic tests. The subjective estimation of the hearing aid user can be elicited with self-assessment inventories encompassing various parameters, e.g., benefit, satisfaction and usage. A questionnaire comprising 11 domains (disability, handicap, frequency and significance of the listening situation, importance of the hearing aid, expectation, demand, aided performance, benefit, satisfaction and usage) within three different conditions (speech in quiet and in noise and listening to sounds) was used to detect components underlying hearing aid fitting. The data show a three-factor structure (situation-, restriction- and aid-related variables) independent from the conditions. Usage depends on all of the three factors. Disability and handicap reveal the highest values for speech in noise, whereas the aid-related factor shows the lowest values for this condition. Global satisfaction with the hearing aid is significantly correlated with the aid-related factor, but independent from the restriction of hearing. The aid-related factor is positively influenced by the amount of social activity because more active persons report higher benefit and satisfaction for all listening conditions. Age does not exhibit a significant relationship to one of the components. Basically, all correlation coefficients are only intermediate, revealing that inter-individual differences of the patients are rather high. The data indicate that extra-audiological factors might also play an important role in the success of hearing aid fitting.
Roverud, Elin; Strickland, Elizabeth A.
2015-01-01
Previous psychoacoustical and physiological studies indicate that the medial olivocochlear reflex (MOCR), a bilateral, sound-evoked reflex, may lead to improved sound intensity discrimination in background noise. The MOCR can decrease the range of basilar-membrane compression and can counteract effects of neural adaptation from background noise. However, the contribution of these processes to intensity discrimination is not well understood. This study examined the effect of ipsilateral, contralateral, and bilateral noise on the “mid-level hump.” The mid-level hump refers to intensity discrimination Weber fractions (WFs) measured for short-duration, high-frequency tones which are poorer at mid levels than at lower or higher levels. The mid-level hump WFs may reflect a limitation due to basilar-membrane compression, and thus may be decreased by the MOCR. The noise was either short (50 ms) or long (150 ms), with the long noise intended to elicit the sluggish MOCR. For a tone in quiet, mid-level hump WFs improved with ipsilateral noise for most listeners, but not with contralateral noise. For a tone in ipsilateral noise, WFs improved with contralateral noise for most listeners, but only when both noises were long. These results are consistent with MOCR-induced WF improvements, possibly via decreases in effects of compression and neural adaptation. PMID:26627798
Tuning in and tuning out: Speech perception in native- and foreign-talker babble
NASA Astrophysics Data System (ADS)
van Heukelem, Kristin; Bradlow, Ann R.
2005-09-01
Studies on speech perception in multitalker babble have revealed asymmetries in the effects of noise on native versus foreign-accented speech intelligibility for native listeners [Rogers et al., Lang Speech 47(2), 139-154 (2004)] and on sentence-in-noise perception by native versus non-native listeners [Mayo et al., J. Speech Lang. Hear. Res., 40, 686-693 (1997)], suggesting that the linguistic backgrounds of talkers and listeners contribute to the effects of noise on speech perception. However, little attention has been paid to the language of the babble. This study tested whether the language of the noise also has asymmetrical effects on listeners. Replicating previous findings [e.g., Bronkhorst and Plomp, J. Acoust. Soc. Am., 92, 3132-3139 (1992)], the results showed poorer English sentence recognition by native English listeners in six-talker babble than in two-talker babble regardless of the language of the babble, demonstrating the effect of increased psychoacoustic/energetic masking. In addition, the results showed that in the two-talker babble condition, native English listeners were more adversely affected by English than Chinese babble. These findings demonstrate informational/cognitive masking on sentence-in-noise recognition in the form of linguistic competition. Whether this competition is at the lexical or sublexical level and whether it is modulated by the phonetic similarity between the target and noise languages remains to be determined.
Effects of linguistic experience on early levels of perceptual tone processing
NASA Astrophysics Data System (ADS)
Huang, Tsan; Johnson, Keith
2005-04-01
This study investigated the phenomenon of language-specificity in Mandarin Chinese tone perception. The main question was whether linguistic experience affects the earliest levels of perceptual processing of tones. Chinese and American English listeners participated in four perception experiments, which involved short inter-stimulus intervals (300 ms or 100 ms) and an AX discrimination or AX degree-of-difference rating task. Three experiments used natural speech monosyllabic tone stimuli and one experiment used time-varying sinusoidal simulations of Mandarin tones. AE listeners showed psychoacoustic listening in all experiments, paying much attention to onset and offset pitch. Chinese listeners showed language-specific patterns in all experiments to various degrees, where tonal neutralization rules reduced perceptual distance between two otherwise contrastive tones for Chinese listeners. Since these experiments employed procedures hypothesized to tap the auditory trace mode (Pisoni, Percept. Psychophys. 13, 253-260 (1973)], language-specificity found in this study seems to support the proposal of an auditory cortical map [Guenther et al., J. Acoust. Soc. Am. 23, 213-221 (1999)]. But the model needs refining to account for different degrees of language-specificity, which are better handled by Johnsons (2004, TLS03:26-41) lexical distance model, although the latter model is too rigid in assuming that linguistic experience does not affect low-level perceptual tasks such as AX discrimination with short ISIs.
Evaluation of central auditory processing in children with Specific Language Impairment.
Włodarczyk, Elżbieta; Szkiełkowska, Agata; Piłka, Adam; Skarżyński, Henryk
2015-01-01
Specific Language Impairment (SLI) affects about 7-15 % of children of school age and according to the currently accepted diagnostic criteria, it is presumed that these children do not suffer from hearing impairment. The goal of this work was to assess anomalies of central auditory processes in a group of children diagnosed with specific language impairment. Material consisted of 200 children aged 7-10 years (100 children in the study group and 100 hundred in the control group). Selected psychoacoustic tests (Frequency Pattern Test - FPT, Duration Pattern Test - DPT, Dichotic Digit Test - DDT, Time Compressed Sentence Test - CST, Gap Detection Test - GDT) were performed in all children. Results were subject to statistical analysis. It was observed that mean results obtained in individual age groups in the study group are significantly lower than in the control group. Based on the conducted studies we may conclude that children with SLI suffer from disorders of some higher auditory functions, which substantiates the diagnosis of hearing disorders according to the AHSA (American Hearing and Speech Association) guidelines. Use of sound-based, not verbal tests, eliminates the probability that observed problems with perception involve only perception of speech, therefore do not signify central hearing disorders, but problems with understanding of speech. Lack of literature data on the significance of FPT, DPT, DDT, CST and GDT tests in children with specific language impairment precludes comparison of acquired results and makes them unique.
Kastelein, Ronald A; Wensveen, Paul J; Terhune, John M; de Jong, Christ A F
2011-01-01
Equal-loudness functions describe relationships between the frequencies of sounds and their perceived loudness. This pilot study investigated the possibility of deriving equal-loudness contours based on the assumption that sounds of equal perceived loudness elicit equal reaction times (RTs). During a psychoacoustic underwater hearing study, the responses of two young female harbor seals to tonal signals between 0.125 and 100 kHz were filmed. Frame-by-frame analysis was used to quantify RT (the time between the onset of the sound stimulus and the onset of movement of the seal away from the listening station). Near-threshold equal-latency contours, as surrogates for equal-loudness contours, were estimated from RT-level functions fitted to mean RT data. The closer the received sound pressure level was to the 50% detection hearing threshold, the more slowly the animals reacted to the signal (RT range: 188-982 ms). Equal-latency contours were calculated relative to the RTs shown by each seal at sound levels of 0, 10, and 20 dB above the detection threshold at 1 kHz. Fifty percent detection thresholds are obtained with well-trained subjects actively listening for faint familiar sounds. When calculating audibility ranges of sounds for harbor seals in nature, it may be appropriate to consider levels 20 dB above this threshold.
Hoare, Derek J; Kowalkowski, Victoria L; Hall, Deborah A
2012-08-01
That auditory perceptual training may alleviate tinnitus draws on two observations: (1) tinnitus probably arises from altered activity within the central auditory system following hearing loss and (2) sound-based training can change central auditory activity. Training that provides sound enrichment across hearing loss frequencies has therefore been hypothesised to alleviate tinnitus. We tested this prediction with two randomised trials of frequency discrimination training involving a total of 70 participants with chronic subjective tinnitus. Participants trained on either (1) a pure-tone standard at a frequency within their region of normal hearing, (2) a pure-tone standard within the region of hearing loss or (3) a high-pass harmonic complex tone spanning a region of hearing loss. Analysis of the primary outcome measure revealed an overall reduction in self-reported tinnitus handicap after training that was maintained at a 1-month follow-up assessment, but there were no significant differences between groups. Secondary analyses also report the effects of different domains of tinnitus handicap on the psychoacoustical characteristics of the tinnitus percept (sensation level, bandwidth and pitch) and on duration of training. Our overall findings and conclusions cast doubt on the superiority of a purely acoustic mechanism to underpin tinnitus remediation. Rather, the nonspecific patterns of improvement are more suggestive that auditory perceptual training affects impact on a contributory mechanism such as selective attention or emotional state.
Neurophysiological model of tinnitus: dependence of the minimal masking level on treatment outcome.
Jastreboff, P J; Hazell, J W; Graham, R L
1994-11-01
Validity of the neurophysiological model of tinnitus (Jastreboff, 1990), outlined in this paper, was tested on data from multicenter trial of tinnitus masking (Hazell et al., 1985). Minimal masking level, intensity match of tinnitus, and the threshold of hearing have been evaluated on a total of 382 patients before and after 6 months of treatment with maskers, hearing aids, or combination devices. The data has been divided into categories depending on treatment outcome and type of approach used. Results of analysis revealed that: i) the psychoacoustical description of tinnitus does not possess a predictive value for the outcome of the treatment; ii) minimal masking level changed significantly depending on the treatment outcome, decreasing on average by 5.3 dB in patients reporting improvement, and increasing by 4.9 dB in those whose tinnitus remained the same or worsened; iii) 73.9% of patients reporting improvement had their minimal masking level decreased as compared with 50.5% for patients not showing improvement, which is at the level of random change; iv) the type of device used has no significant impact on the treatment outcome and minimal masking level change; v) intensity match and threshold of hearing did not exhibit any significant changes which can be related to treatment outcome. These results are fully consistent with the neurophysiological interpretation of mechanisms involved in the phenomenon of tinnitus and its alleviation.
Gee, Bryan M; Thompson, Kelly; St John, Holly
2014-03-01
Sound-based interventions (SBIs) are being used by paediatric occupational therapists to help children with autism spectrum disorders and co-morbid sensory processing disorders. A limited yet growing body of evidence is emerging related to the efficacy of SBIs in reducing sensory processing deficits among paediatric clients with co-morbid conditions. The current study employed an ABA single-subject case-controlled design, implementing The Listening Program® with a 7-year-old child diagnosed with autism spectrum disorder who demonstrated auditory sensory over-responsivity (SOR). The intervention consisted of 10 weeks of psycho-acoustically modified classical music that was delivered using specialized headphones and amplifier and a standard CD player. Repeated measures were conducted during the A(1), B and A(2) phases of the study using the Sensory Processing Measure, a subjective caregiver questionnaire, and the Sensory Over-Responsivity Scales, an examiner-based assessment measure to track changes of the participant's auditory SOR-related behaviours. The results indicated that the participant exhibited a decrease in the number of negative (avoidant, verbal and physical negative) and self-stimulatory behaviours. The decreases in negative and self-stimulatory behaviour may have been due to the therapeutic effect of the repeated exposure to the Sensory Over-Responsivity Scales or The Listening Program SBI. Copyright © 2013 John Wiley & Sons, Ltd.
Quiet as an Environmental Value: A Contrast between Two Legislative Approaches
Thorne, Robert; Shepherd, Daniel
2013-01-01
This paper examines the concept of “quiet” as an “environmental value” in terms of amenity and wellbeing from a legislative context. Critical review of two pieces of environmental legislation from Australia and New Zealand forms the basis of the paper. The Australian legislation is Queensland’s Environmental Protection Act, and the New Zealand legislation is that nation’s Resource Management Act. Quiet is part of the psychoacoustic continuum between a tranquil and an intrusively noisy sound environment. As such, quiet possesses intrinsic value in terms of overall sound within the environment (soundscape) and to individuals and communities. In both pieces of legislation, guidance, either directly or indirectly, is given to “maximum” sound levels to describe the acoustic environment. Only in Queensland is wellbeing and amenity described as environmental values, while in the New Zealand approach, amenity is identified as the core value to defend, but guidance is not well established. Wellbeing can be related to degrees of quietness and the absence of intrusive noise, the character of sound within an environment (“soundscape”), as well as the overall level of sound. The quality of life experienced by individuals is related to that person’s physical and mental health, sense of amenity and wellbeing. These characteristics can be described in terms of subjective and objective measures, though legislation does not always acknowledge the subjective. PMID:23823712
On the Etiology of Listening Difficulties in Noise Despite Clinically Normal Audiograms
2017-01-01
Many people with difficulties following conversations in noisy settings have “clinically normal” audiograms, that is, tone thresholds better than 20 dB HL from 0.1 to 8 kHz. This review summarizes the possible causes of such difficulties, and examines established as well as promising new psychoacoustic and electrophysiologic approaches to differentiate between them. Deficits at the level of the auditory periphery are possible even if thresholds remain around 0 dB HL, and become probable when they reach 10 to 20 dB HL. Extending the audiogram beyond 8 kHz can identify early signs of noise-induced trauma to the vulnerable basal turn of the cochlea, and might point to “hidden” losses at lower frequencies that could compromise speech reception in noise. Listening difficulties can also be a consequence of impaired central auditory processing, resulting from lesions affecting the auditory brainstem or cortex, or from abnormal patterns of sound input during developmental sensitive periods and even in adulthood. Such auditory processing disorders should be distinguished from (cognitive) linguistic deficits, and from problems with attention or working memory that may not be specific to the auditory modality. Improved diagnosis of the causes of listening difficulties in noise should lead to better treatment outcomes, by optimizing auditory training procedures to the specific deficits of individual patients, for example. PMID:28002080
The effects of noise on the cognitive performance of physicians in a hospital emergency department
NASA Astrophysics Data System (ADS)
Dodds, Peter
In this research, the acoustic environment of a contemporary urban hospital emergency department has been characterized. Perceptive and cognitive tests relating to the acoustic environment were conducted on both medical professionals and lay people and a methodology for developing augmentable acoustic simulations from field recordings was developed. While research of healthcare environments remains a popular area of investigation for the acoustics community, a lack of communication between medical and acoustics researchers as well as a lack of sophistication in the methods implemented to evaluate hospital environments and their occupants has led to stagnation. This research attempted to replicate traditional methods for the evaluation of hospital acoustic environments including impulse response based room acoustics measurements as well as psychoacoustic evaluations. This thesis also demonstrates some of the issues associated with conducting such research and provides an outline and implementation for alternative advanced methods of re- search. Advancements include the use of the n-Back test to evaluate the effects of the acoustic environment on cognitive function as well as the outline of a new methodology for implementing realistic immersive simulations for cognitive and perceptual testing using field recordings and signal processing techniques. Additionally, this research utilizes feedback from working emergency medicine physicians to determine the subjective degree of distraction subjects felt in response to a simulated acoustic environment. Results of the room acoustics measurements and all experiments will be presented and analyzed and possible directions for future research will be presented.
Is There a Right Ear Advantage in Congenital Aural Atresia?
Reed, Robert; Hubbard, Matthew; Kesser, Bradley W
2016-12-01
To compare speech/language development and academic progress between children with right versus left congenital aural atresia (CAA). Case control survey and review of audiometric data. Tertiary care academic practice. Children with unilateral CAA. Demographic and audiometric data; rates of grade retention, use of any hearing or learning resource, and behavioral problems. No significant differences in grade retention rate, utilization of amplification, speech language therapy, use of an individualized education program, or frequency modulated system were found between children with right versus left CAA. Children with left CAA were significantly more likely to be enrolled in special education programs (p = 0.026). Differences in reported communication problems approached significance with more difficulty noted in the right ear group (p = 0.059). Left CAA patients were also more likely to have reported behavioral problems (p = 0.0039). Contrary to the hypothesis that a normal hearing right ear confers a language advantage in patients with unilateral hearing loss, children with left CAA (normal right ear) were statistically more likely to be enrolled in a special education program and have behavioral problems. Reported communication problems were more common in right CAA patients, but this did not reach statistical significance. No differences were found in use of amplification, frequency modulated system, individualized education program, or grade retention. Further investigation of both the clinical implications and underlying psychoacoustics of unilateral hearing loss and the identification and habilitation of "at risk" unilateral hearing loss children is warranted.
Zipf's Law in Short-Time Timbral Codings of Speech, Music, and Environmental Sound Signals
Haro, Martín; Serrà, Joan; Herrera, Perfecto; Corral, Álvaro
2012-01-01
Timbre is a key perceptual feature that allows discrimination between different sounds. Timbral sensations are highly dependent on the temporal evolution of the power spectrum of an audio signal. In order to quantitatively characterize such sensations, the shape of the power spectrum has to be encoded in a way that preserves certain physical and perceptual properties. Therefore, it is common practice to encode short-time power spectra using psychoacoustical frequency scales. In this paper, we study and characterize the statistical properties of such encodings, here called timbral code-words. In particular, we report on rank-frequency distributions of timbral code-words extracted from 740 hours of audio coming from disparate sources such as speech, music, and environmental sounds. Analogously to text corpora, we find a heavy-tailed Zipfian distribution with exponent close to one. Importantly, this distribution is found independently of different encoding decisions and regardless of the audio source. Further analysis on the intrinsic characteristics of most and least frequent code-words reveals that the most frequent code-words tend to have a more homogeneous structure. We also find that speech and music databases have specific, distinctive code-words while, in the case of the environmental sounds, this database-specific code-words are not present. Finally, we find that a Yule-Simon process with memory provides a reasonable quantitative approximation for our data, suggesting the existence of a common simple generative mechanism for all considered sound sources. PMID:22479497
Rationale for the tinnitus retraining therapy trial.
Formby, Craig; Scherer, Roberta
2013-01-01
The Tinnitus Retraining Therapy Trial (TRTT) is a National Institutes of Health-sponsored, multi-centered, placebo-controlled, randomized trial evaluating the efficacy of tinnitus retraining therapy (TRT) and its component parts, directive counseling and sound therapy, as treatments for subjective debilitating tinnitus in the military. The TRTT will enroll 228 individuals at an allocation ratio of 1:1:1 to: (1) directive counseling and sound therapy using conventional sound generators; (2) directive counseling and placebo sound generators; or (3) standard of care as administered in the military. Study centers include a Study Chair's Office, a Data Coordinating Center, and six Military Clinical Centers with treatment and data collection standardized across all clinics. The primary outcome is change in Tinnitus Questionnaire (TQ) score assessed longitudinally at 3, 6, 12, and 18-month follow-up visits. Secondary outcomes include: Change in TQ sub-scales, Tinnitus Handicap Inventory, Tinnitus Functional Index, and TRT interview visual analog scale; audiometric and psychoacoustic measures; and change in quality of life. The TRTT will evaluate TRT efficacy by comparing TRT (directive counseling and conventional sound generators) with standard of care; directive counseling by comparing directive counseling plus placebo sound generators versus standard of care; and sound therapy by comparing conventional versus placebo sound generators. We hypothesize that full TRT will be more efficacious than standard of care, directive counseling and placebo sound generators more efficacious than standard of care, and conventional more efficacious than placebo sound generators in habituating the tinnitus awareness, annoyance, and impact on the study participant's life.
Schädler, Marc René; Warzybok, Anna; Meyer, Bernd T.; Brand, Thomas
2016-01-01
To characterize the individual patient’s hearing impairment as obtained with the matrix sentence recognition test, a simulation Framework for Auditory Discrimination Experiments (FADE) is extended here using the Attenuation and Distortion (A+D) approach by Plomp as a blueprint for setting the individual processing parameters. FADE has been shown to predict the outcome of both speech recognition tests and psychoacoustic experiments based on simulations using an automatic speech recognition system requiring only few assumptions. It builds on the closed-set matrix sentence recognition test which is advantageous for testing individual speech recognition in a way comparable across languages. Individual predictions of speech recognition thresholds in stationary and in fluctuating noise were derived using the audiogram and an estimate of the internal level uncertainty for modeling the individual Plomp curves fitted to the data with the Attenuation (A-) and Distortion (D-) parameters of the Plomp approach. The “typical” audiogram shapes from Bisgaard et al with or without a “typical” level uncertainty and the individual data were used for individual predictions. As a result, the individualization of the level uncertainty was found to be more important than the exact shape of the individual audiogram to accurately model the outcome of the German Matrix test in stationary or fluctuating noise for listeners with hearing impairment. The prediction accuracy of the individualized approach also outperforms the (modified) Speech Intelligibility Index approach which is based on the individual threshold data only. PMID:27604782
The Gap Detection Test: Can It Be Used to Diagnose Tinnitus?
Boyen, Kris; Başkent, Deniz; van Dijk, Pim
2015-01-01
Animals with induced tinnitus showed difficulties in detecting silent gaps in sounds, suggesting that the tinnitus percept may be filling the gap. The main purpose of this study was to evaluate the applicability of this approach to detect tinnitus in human patients. The authors first hypothesized that gap detection would be impaired in patients with tinnitus, and second, that gap detection would be more impaired at frequencies close to the tinnitus frequency of the patient. Twenty-two adults with bilateral tinnitus, 20 age-matched and hearing loss-matched subjects without tinnitus, and 10 young normal-hearing subjects participated in the study. To determine the characteristics of the tinnitus, subjects matched an external sound to their perceived tinnitus in pitch and loudness. To determine the minimum detectable gap, the gap threshold, an adaptive psychoacoustic test was performed three times by each subject. In this gap detection test, four different stimuli, with various frequencies and bandwidths, were presented at three intensity levels each. Similar to previous reports of gap detection, increasing sensation level yielded shorter gap thresholds for all stimuli in all groups. Interestingly, the tinnitus group did not display elevated gap thresholds in any of the four stimuli. Moreover, visual inspection of the data revealed no relation between gap detection performance and perceived tinnitus pitch. These findings show that tinnitus in humans has no effect on the ability to detect gaps in auditory stimuli. Thus, the testing procedure in its present form is not suitable for clinical detection of tinnitus in humans.
Grimm, Giso; Hohmann, Volker; Laugesen, Søren; Neher, Tobias
2017-01-01
In contrast to static sounds, spatially dynamic sounds have received little attention in psychoacoustic research so far. This holds true especially for acoustically complex (reverberant, multisource) conditions and impaired hearing. The current study therefore investigated the influence of reverberation and the number of concurrent sound sources on source movement detection in young normal-hearing (YNH) and elderly hearing-impaired (EHI) listeners. A listening environment based on natural environmental sounds was simulated using virtual acoustics and rendered over headphones. Both near-far (‘radial’) and left-right (‘angular’) movements of a frontal target source were considered. The acoustic complexity was varied by adding static lateral distractor sound sources as well as reverberation. Acoustic analyses confirmed the expected changes in stimulus features that are thought to underlie radial and angular source movements under anechoic conditions and suggested a special role of monaural spectral changes under reverberant conditions. Analyses of the detection thresholds showed that, with the exception of the single-source scenarios, the EHI group was less sensitive to source movements than the YNH group, despite adequate stimulus audibility. Adding static sound sources clearly impaired the detectability of angular source movements for the EHI (but not the YNH) group. Reverberation, on the other hand, clearly impaired radial source movement detection for the EHI (but not the YNH) listeners. These results illustrate the feasibility of studying factors related to auditory movement perception with the help of the developed test setup. PMID:28675088
Kastelein, Ronald A; Wensveen, Paul; Hoek, Lean; Terhune, John M
2009-07-01
The underwater hearing sensitivities of two 1.5-year-old female harbor seals were quantified in a quiet pool built specifically for acoustic research, by using a behavioral psychoacoustic technique. The animals were trained to respond when they detected an acoustic signal and not to respond when they did not ("go/no-go" response). Fourteen narrowband noise signals (1/3-octave bands but with some energy in adjacent bands), at 1/3-octave center frequencies of 0.2-80 kHz, and of 900 ms duration, were tested. Thresholds at each frequency were measured using the up-down staircase method and defined as the stimulus level resulting in a 50% detection rate. Between 0.5 and 40 kHz, the thresholds corresponded to a 1/3-octave band noise level of approximately 60 dB re 1 microPa (SD+/-3.0 dB). At lower frequencies, the thresholds increased to 66 dB re 1 microPa and at 80 kHz the thresholds rose to 114 dB re 1 microPa. The 1/3-octave noise band thresholds of the two seals did not differ from each other, or from the narrowband frequency-modulated tone thresholds at the same frequencies obtained a few months before for the same animals. These hearing threshold values can be used to calculate detection ranges of underwater calls and anthropogenic noises by harbor seals.
Underwater detection of tonal signals between 0.125 and 100 kHz by harbor seals (Phoca vitulina).
Kastelein, Ronald A; Wensveen, Paul J; Hoek, Lean; Verboom, Willem C; Terhune, John M
2009-02-01
The underwater hearing sensitivities of two 1-year-old female harbor seals were quantified in a pool built for acoustic research, using a behavioral psychoacoustic technique. The animals were trained to respond when they detected an acoustic signal and not to respond when they did not (go/no-go response). Pure tones (0.125-0.25 kHz) and narrowband frequency modulated (tonal) signals (center frequencies 0.5-100 kHz) of 900 ms duration were tested. Thresholds at each frequency were measured using the up-down staircase method and defined as the stimulus level resulting in a 50% detection rate. The audiograms of the two seals did not differ statistically: both plots showed the typical mammalian U-shape, but with a wide and flat bottom. Maximum sensitivity (54 dB re 1 microPa, rms) occurred at 1 kHz. The frequency range of best hearing (within 10 dB of maximum sensitivity) was from 0.5 to 40 kHz (6(1/3) octaves). Higher hearing thresholds (indicating poorer sensitivity) were observed below 1 and above 40 kHz. Thresholds below 4 kHz were lower than those previously described for harbor seals, which demonstrates the importance of using quiet facilities, built specifically for acoustic research, for hearing studies in marine mammals. The results suggest that under unmasked conditions many anthropogenic noise sources and sounds from conspecifics are audible to harbor seals at greater ranges than formerly believed.
Leftward lateralization of auditory cortex underlies holistic sound perception in Williams syndrome.
Wengenroth, Martina; Blatow, Maria; Bendszus, Martin; Schneider, Peter
2010-08-23
Individuals with the rare genetic disorder Williams-Beuren syndrome (WS) are known for their characteristic auditory phenotype including strong affinity to music and sounds. In this work we attempted to pinpoint a neural substrate for the characteristic musicality in WS individuals by studying the structure-function relationship of their auditory cortex. Since WS subjects had only minor musical training due to psychomotor constraints we hypothesized that any changes compared to the control group would reflect the contribution of genetic factors to auditory processing and musicality. Using psychoacoustics, magnetoencephalography and magnetic resonance imaging, we show that WS individuals exhibit extreme and almost exclusive holistic sound perception, which stands in marked contrast to the even distribution of this trait in the general population. Functionally, this was reflected by increased amplitudes of left auditory evoked fields. On the structural level, volume of the left auditory cortex was 2.2-fold increased in WS subjects as compared to control subjects. Equivalent volumes of the auditory cortex have been previously reported for professional musicians. There has been an ongoing debate in the neuroscience community as to whether increased gray matter of the auditory cortex in musicians is attributable to the amount of training or innate disposition. In this study musical education of WS subjects was negligible and control subjects were carefully matched for this parameter. Therefore our results not only unravel the neural substrate for this particular auditory phenotype, but in addition propose WS as a unique genetic model for training-independent auditory system properties.
Lundbeck, Micha; Grimm, Giso; Hohmann, Volker; Laugesen, Søren; Neher, Tobias
2017-01-01
In contrast to static sounds, spatially dynamic sounds have received little attention in psychoacoustic research so far. This holds true especially for acoustically complex (reverberant, multisource) conditions and impaired hearing. The current study therefore investigated the influence of reverberation and the number of concurrent sound sources on source movement detection in young normal-hearing (YNH) and elderly hearing-impaired (EHI) listeners. A listening environment based on natural environmental sounds was simulated using virtual acoustics and rendered over headphones. Both near-far ('radial') and left-right ('angular') movements of a frontal target source were considered. The acoustic complexity was varied by adding static lateral distractor sound sources as well as reverberation. Acoustic analyses confirmed the expected changes in stimulus features that are thought to underlie radial and angular source movements under anechoic conditions and suggested a special role of monaural spectral changes under reverberant conditions. Analyses of the detection thresholds showed that, with the exception of the single-source scenarios, the EHI group was less sensitive to source movements than the YNH group, despite adequate stimulus audibility. Adding static sound sources clearly impaired the detectability of angular source movements for the EHI (but not the YNH) group. Reverberation, on the other hand, clearly impaired radial source movement detection for the EHI (but not the YNH) listeners. These results illustrate the feasibility of studying factors related to auditory movement perception with the help of the developed test setup.
Grose, John H; Buss, Emily; Hall, Joseph W
2017-01-01
The purpose of this study was to test the hypothesis that listeners with frequent exposure to loud music exhibit deficits in suprathreshold auditory performance consistent with cochlear synaptopathy. Young adults with normal audiograms were recruited who either did ( n = 31) or did not ( n = 30) have a history of frequent attendance at loud music venues where the typical sound levels could be expected to result in temporary threshold shifts. A test battery was administered that comprised three sets of procedures: (a) electrophysiological tests including distortion product otoacoustic emissions, auditory brainstem responses, envelope following responses, and the acoustic change complex evoked by an interaural phase inversion; (b) psychoacoustic tests including temporal modulation detection, spectral modulation detection, and sensitivity to interaural phase; and (c) speech tests including filtered phoneme recognition and speech-in-noise recognition. The results demonstrated that a history of loud music exposure can lead to a profile of peripheral auditory function that is consistent with an interpretation of cochlear synaptopathy in humans, namely, modestly abnormal auditory brainstem response Wave I/Wave V ratios in the presence of normal distortion product otoacoustic emissions and normal audiometric thresholds. However, there were no other electrophysiological, psychophysical, or speech perception effects. The absence of any behavioral effects in suprathreshold sound processing indicated that, even if cochlear synaptopathy is a valid pathophysiological condition in humans, its perceptual sequelae are either too diffuse or too inconsequential to permit a simple differential diagnosis of hidden hearing loss.
Bailes, Freya; Dean, Roger T; Broughton, Mary C
2015-01-01
For listeners familiar with Western twelve-tone equal-tempered (12-TET) music, a novel microtonal tuning system is expected to present additional processing challenges. We aimed to determine whether this was the case, focusing on the extent to which our perceptions can be considered bottom-up (psychoacoustic and primarily perceptual) and top-down (dependent on familiarity and cognitive processing). We elicited both overt response ratings, and covert event-related potentials (ERPs), so as to compare subjective impressions of sounds with the neurophysiological processing of the acoustic signal. We hypothesised that microtonal intervals are perceived differently from 12-TET intervals, and that the responses of musicians (n = 10) and non-musicians (n = 10) are distinct. Two-note chords were presented comprising 12-TET intervals (consonant and dissonant) or microtonal (quarter tone) intervals, and ERP, subjective roughness ratings, and liking ratings were recorded successively. Musical experience mediated the perception of differences between dissonant and microtone intervals, with non-musicians giving similar ratings for each, and musicians preferring dissonant over the less commonly used microtonal intervals, rating them as less rough. ERP response amplitude was greater for consonant intervals than other intervals. Musical experience interacted with interval type, suggesting that musical expertise facilitates the sensory and perceptual discrimination of microtonal intervals from 12-TET intervals, and an increased ability to categorize such intervals. Non-musicians appear to have perceived microtonal intervals as instances of neighbouring 12-TET intervals.
Music and Language Syntax Interact in Broca's Area: An fMRI Study.
Kunert, Richard; Willems, Roel M; Casasanto, Daniel; Patel, Aniruddh D; Hagoort, Peter
2015-01-01
Instrumental music and language are both syntactic systems, employing complex, hierarchically-structured sequences built using implicit structural norms. This organization allows listeners to understand the role of individual words or tones in the context of an unfolding sentence or melody. Previous studies suggest that the brain mechanisms of syntactic processing may be partly shared between music and language. However, functional neuroimaging evidence for anatomical overlap of brain activity involved in linguistic and musical syntactic processing has been lacking. In the present study we used functional magnetic resonance imaging (fMRI) in conjunction with an interference paradigm based on sung sentences. We show that the processing demands of musical syntax (harmony) and language syntax interact in Broca's area in the left inferior frontal gyrus (without leading to music and language main effects). A language main effect in Broca's area only emerged in the complex music harmony condition, suggesting that (with our stimuli and tasks) a language effect only becomes visible under conditions of increased demands on shared neural resources. In contrast to previous studies, our design allows us to rule out that the observed neural interaction is due to: (1) general attention mechanisms, as a psychoacoustic auditory anomaly behaved unlike the harmonic manipulation, (2) error processing, as the language and the music stimuli contained no structural errors. The current results thus suggest that two different cognitive domains-music and language-might draw on the same high level syntactic integration resources in Broca's area.
Neural Biomarkers for Dyslexia, ADHD, and ADD in the Auditory Cortex of Children.
Serrallach, Bettina; Groß, Christine; Bernhofs, Valdis; Engelmann, Dorte; Benner, Jan; Gündert, Nadine; Blatow, Maria; Wengenroth, Martina; Seitz, Angelika; Brunner, Monika; Seither, Stefan; Parncutt, Richard; Schneider, Peter; Seither-Preisler, Annemarie
2016-01-01
Dyslexia, attention deficit hyperactivity disorder (ADHD), and attention deficit disorder (ADD) show distinct clinical profiles that may include auditory and language-related impairments. Currently, an objective brain-based diagnosis of these developmental disorders is still unavailable. We investigated the neuro-auditory systems of dyslexic, ADHD, ADD, and age-matched control children (N = 147) using neuroimaging, magnetencephalography and psychoacoustics. All disorder subgroups exhibited an oversized left planum temporale and an abnormal interhemispheric asynchrony (10-40 ms) of the primary auditory evoked P1-response. Considering right auditory cortex morphology, bilateral P1 source waveform shapes, and auditory performance, the three disorder subgroups could be reliably differentiated with outstanding accuracies of 89-98%. We therefore for the first time provide differential biomarkers for a brain-based diagnosis of dyslexia, ADHD, and ADD. The method allowed not only allowed for clear discrimination between two subtypes of attentional disorders (ADHD and ADD), a topic controversially discussed for decades in the scientific community, but also revealed the potential for objectively identifying comorbid cases. Noteworthy, in children playing a musical instrument, after three and a half years of training the observed interhemispheric asynchronies were reduced by about 2/3, thus suggesting a strong beneficial influence of music experience on brain development. These findings might have far-reaching implications for both research and practice and enable a profound understanding of the brain-related etiology, diagnosis, and musically based therapy of common auditory-related developmental disorders and learning disabilities.
Neural Biomarkers for Dyslexia, ADHD, and ADD in the Auditory Cortex of Children
Serrallach, Bettina; Groß, Christine; Bernhofs, Valdis; Engelmann, Dorte; Benner, Jan; Gündert, Nadine; Blatow, Maria; Wengenroth, Martina; Seitz, Angelika; Brunner, Monika; Seither, Stefan; Parncutt, Richard; Schneider, Peter; Seither-Preisler, Annemarie
2016-01-01
Dyslexia, attention deficit hyperactivity disorder (ADHD), and attention deficit disorder (ADD) show distinct clinical profiles that may include auditory and language-related impairments. Currently, an objective brain-based diagnosis of these developmental disorders is still unavailable. We investigated the neuro-auditory systems of dyslexic, ADHD, ADD, and age-matched control children (N = 147) using neuroimaging, magnetencephalography and psychoacoustics. All disorder subgroups exhibited an oversized left planum temporale and an abnormal interhemispheric asynchrony (10–40 ms) of the primary auditory evoked P1-response. Considering right auditory cortex morphology, bilateral P1 source waveform shapes, and auditory performance, the three disorder subgroups could be reliably differentiated with outstanding accuracies of 89–98%. We therefore for the first time provide differential biomarkers for a brain-based diagnosis of dyslexia, ADHD, and ADD. The method allowed not only allowed for clear discrimination between two subtypes of attentional disorders (ADHD and ADD), a topic controversially discussed for decades in the scientific community, but also revealed the potential for objectively identifying comorbid cases. Noteworthy, in children playing a musical instrument, after three and a half years of training the observed interhemispheric asynchronies were reduced by about 2/3, thus suggesting a strong beneficial influence of music experience on brain development. These findings might have far-reaching implications for both research and practice and enable a profound understanding of the brain-related etiology, diagnosis, and musically based therapy of common auditory-related developmental disorders and learning disabilities. PMID:27471442
Kwon, Bomjun J
2012-06-01
This article introduces AUX (AUditory syntaX), a scripting syntax specifically designed to describe auditory signals and processing, to the members of the behavioral research community. The syntax is based on descriptive function names and intuitive operators suitable for researchers and students without substantial training in programming, who wish to generate and examine sound signals using a written script. In this article, the essence of AUX is discussed and practical examples of AUX scripts specifying various signals are illustrated. Additionally, two accompanying Windows-based programs and development libraries are described. AUX Viewer is a program that generates, visualizes, and plays sounds specified in AUX. AUX Viewer can also be used for class demonstrations or presentations. Another program, Psycon, allows a wide range of sound signals to be used as stimuli in common psychophysical testing paradigms, such as the adaptive procedure, the method of constant stimuli, and the method of adjustment. AUX Library is also provided, so that researchers can develop their own programs utilizing AUX. The philosophical basis of AUX is to separate signal generation from the user interface needed for experiments. AUX scripts are portable and reusable; they can be shared by other researchers, regardless of differences in actual AUX-based programs, and reused for future experiments. In short, the use of AUX can be potentially beneficial to all members of the research community-both those with programming backgrounds and those without.
Perceptual analyses of spasmodic dysphonia before and after treatment.
Cannito, Michael P; Woodson, Gayle E; Murry, Thomas; Bender, Brenda
2004-12-01
To evaluate expert listeners' perceptions of voice and fluency in persons with adductor spasmodic dysphonia (ADSD) before and after treatment with botulinum toxin type A (Botox), as a function of initial severity of the disorder (while controlling for patients' age at injection). Simple before-and-after trial with blinded randomized listener judgments. Ambulatory care clinic at a single medical center. Forty-two consecutive patients with ADSD who underwent examination, with a 3- to 6-week follow-up, after initial botulinum toxin type A injection. There were also 42 age- and sex-matched healthy control subjects. Injections of botulinum toxin type A into the thyroarytenoid muscle(s). Computer-implemented visual analog scaling judgments of voice quality and speech fluency made by expert listeners under psychoacoustically controlled conditions. Response to botulinum toxin type A varied markedly as a function of pretreatment severity of ADSD. More severe initial symptoms exhibited greater magnitudes of improvement. Patients with mild dysphonia did not exhibit pretreatment to posttreatment change. Following treatment, voice and fluency remained significantly (P<.05) poorer in ADSD than in healthy speakers. Older patients exhibited less improvement than younger patients when the effect of initial severity was statistically controlled. Voice quality and fluency improved for most patients following treatment, but older patients and those with milder dysphonia exhibited the least optimal responses to the procedure. Patients who were profoundly impaired demonstrated the greatest amount of improvement. Computer-implemented visual analog scaling provided a reliable clinical tool for determining treatment-related changes in those with ADSD.
Dental Environmental Noise Evaluation and Health Risk Model Construction to Dental Professionals.
Ma, Kuen Wai; Wong, Hai Ming; Mak, Cheuk Ming
2017-09-19
Occupational noise is unavoidably produced from dental equipment, building facilities, and human voices in the dental environment. The purpose of this study was to investigate the effect of occupational noise exposure on the dental professionals' health condition. The psychoacoustics approach noise exposure assessment followed by the health risk assessment was carried on at the paediatric dentistry clinic and the dental laboratory in the Prince Philip Dental Hospital of Hong Kong. The A-weighted equivalent sound level, total loudness, and sharpness values were statistically significantly higher for the noise at the laboratory than that at the clinic. The degree of perceived influences and sharpness of noise were found to have the impacts on the dental professionals' working performance and health. Moreover, the risk of having a bad hearing state would a have 26% and 31% higher chance for a unit increment of the short-term and long-term impact scores, respectively. The dental professionals with the service length more than 10 years and the daily working hours of more than eight showed the highest risk to their hearing state. The worse the hearing state was, the worse the health state was found for the dental professionals. Also, the risk of dissatisfaction would be increased by 4.41 and 1.22 times for those who worked at the laboratory and a unit increment of the long-term impact score. The constructed health risk mode with the scientific and statistical evidence is hence important for the future noise management of environmental improvement.
Music and Language Syntax Interact in Broca’s Area: An fMRI Study
Kunert, Richard; Willems, Roel M.; Casasanto, Daniel; Patel, Aniruddh D.; Hagoort, Peter
2015-01-01
Instrumental music and language are both syntactic systems, employing complex, hierarchically-structured sequences built using implicit structural norms. This organization allows listeners to understand the role of individual words or tones in the context of an unfolding sentence or melody. Previous studies suggest that the brain mechanisms of syntactic processing may be partly shared between music and language. However, functional neuroimaging evidence for anatomical overlap of brain activity involved in linguistic and musical syntactic processing has been lacking. In the present study we used functional magnetic resonance imaging (fMRI) in conjunction with an interference paradigm based on sung sentences. We show that the processing demands of musical syntax (harmony) and language syntax interact in Broca’s area in the left inferior frontal gyrus (without leading to music and language main effects). A language main effect in Broca’s area only emerged in the complex music harmony condition, suggesting that (with our stimuli and tasks) a language effect only becomes visible under conditions of increased demands on shared neural resources. In contrast to previous studies, our design allows us to rule out that the observed neural interaction is due to: (1) general attention mechanisms, as a psychoacoustic auditory anomaly behaved unlike the harmonic manipulation, (2) error processing, as the language and the music stimuli contained no structural errors. The current results thus suggest that two different cognitive domains—music and language—might draw on the same high level syntactic integration resources in Broca’s area. PMID:26536026
Auditory perception bias in speech imitation
Postma-Nilsenová, Marie; Postma, Eric
2013-01-01
In an experimental study, we explored the role of auditory perception bias in vocal pitch imitation. Psychoacoustic tasks involving a missing fundamental indicate that some listeners are attuned to the relationship between all the higher harmonics present in the signal, which supports their perception of the fundamental frequency (the primary acoustic correlate of pitch). Other listeners focus on the lowest harmonic constituents of the complex sound signal which may hamper the perception of the fundamental. These two listener types are referred to as fundamental and spectral listeners, respectively. We hypothesized that the individual differences in speakers' capacity to imitate F0 found in earlier studies, may at least partly be due to the capacity to extract information about F0 from the speech signal. Participants' auditory perception bias was determined with a standard missing fundamental perceptual test. Subsequently, speech data were collected in a shadowing task with two conditions, one with a full speech signal and one with high-pass filtered speech above 300 Hz. The results showed that perception bias toward fundamental frequency was related to the degree of F0 imitation. The effect was stronger in the condition with high-pass filtered speech. The experimental outcomes suggest advantages for fundamental listeners in communicative situations where F0 imitation is used as a behavioral cue. Future research needs to determine to what extent auditory perception bias may be related to other individual properties known to improve imitation, such as phonetic talent. PMID:24204361
Microphone Handling Noise: Measurements of Perceptual Threshold and Effects on Audio Quality
Kendrick, Paul; Jackson, Iain R.; Fazenda, Bruno M.; Cox, Trevor J.; Li, Francis F.
2015-01-01
A psychoacoustic experiment was carried out to test the effects of microphone handling noise on perceived audio quality. Handling noise is a problem affecting both amateurs using their smartphones and cameras, as well as professionals using separate microphones and digital recorders. The noises used for the tests were measured from a variety of devices, including smartphones, laptops and handheld microphones. The signal features that characterise these noises are analysed and presented. The sounds include various types of transient, impact noises created by tapping or knocking devices, as well as more sustained sounds caused by rubbing. During the perceptual tests, listeners auditioned speech podcasts and were asked to rate the degradation of any unwanted sounds they heard. A representative design test methodology was developed that tried to encourage everyday rather than analytical listening. Signal-to-noise ratio (SNR) of the handling noise events was shown to be the best predictor of quality degradation. Other factors such as noise type or background noise in the listening environment did not significantly affect quality ratings. Podcast, microphone type and reproduction equipment were found to be significant but only to a small extent. A model allowing the prediction of degradation from the SNR is presented. The SNR threshold at which 50% of subjects noticed handling noise was found to be 4.2 ± 0.6 dBA. The results from this work are important for the understanding of our perception of impact sound and resonant noises in recordings, and will inform the future development of an automated predictor of quality for handling noise. PMID:26473498
Residual Inhibition Functions Overlap Tinnitus Spectra and the Region of Auditory Threshold Shift
Moffat, Graeme; Baumann, Michael; Ward, Lawrence M.
2008-01-01
Animals exposed to noise trauma show augmented synchronous neural activity in tonotopically reorganized primary auditory cortex consequent on hearing loss. Diminished intracortical inhibition in the reorganized region appears to enable synchronous network activity that develops when deafferented neurons begin to respond to input via their lateral connections. In humans with tinnitus accompanied by hearing loss, this process may generate a phantom sound that is perceived in accordance with the location of the affected neurons in the cortical place map. The neural synchrony hypothesis predicts that tinnitus spectra, and heretofore unmeasured “residual inhibition functions” that relate residual tinnitus suppression to the center frequency of masking sounds, should cover the region of hearing loss in the audiogram. We confirmed these predictions in two independent cohorts totaling 90 tinnitus subjects, using computer-based tools designed to assess the psychoacoustic properties of tinnitus. Tinnitus spectra and residual inhibition functions for depth and duration increased with the amount of threshold shift over the region of hearing impairment. Residual inhibition depth was shallower when the masking sounds that were used to induce residual inhibition showed decreased correspondence with the frequency spectrum and bandwidth of the tinnitus. These findings suggest that tinnitus and its suppression in residual inhibition depend on processes that span the region of hearing impairment and not on mechanisms that enhance cortical representations for sound frequencies at the audiometric edge. Hearing thresholds measured in age-matched control subjects without tinnitus implicated hearing loss as a factor in tinnitus, although elevated thresholds alone were not sufficient to cause tinnitus. PMID:18712566
Addis, L; Friederici, A D; Kotz, S A; Sabisch, B; Barry, J; Richter, N; Ludwig, A A; Rübsamen, R; Albert, F W; Pääbo, S; Newbury, D F; Monaco, A P
2010-01-01
Despite the apparent robustness of language learning in humans, a large number of children still fail to develop appropriate language skills despite adequate means and opportunity. Most cases of language impairment have a complex etiology, with genetic and environmental influences. In contrast, we describe a three-generation German family who present with an apparently simple segregation of language impairment. Investigations of the family indicate auditory processing difficulties as a core deficit. Affected members performed poorly on a nonword repetition task and present with communication impairments. The brain activation pattern for syllable duration as measured by event-related brain potentials showed clear differences between affected family members and controls, with only affected members displaying a late discrimination negativity. In conjunction with psychoacoustic data showing deficiencies in auditory duration discrimination, the present results indicate increased processing demands in discriminating syllables of different duration. This, we argue, forms the cognitive basis of the observed language impairment in this family. Genome-wide linkage analysis showed a haplotype in the central region of chromosome 12 which reaches the maximum possible logarithm of odds ratio (LOD) score and fully co-segregates with the language impairment, consistent with an autosomal dominant, fully penetrant mode of inheritance. Whole genome analysis yielded no novel inherited copy number variants strengthening the case for a simple inheritance pattern. Several genes in this region of chromosome 12 which are potentially implicated in language impairment did not contain polymorphisms likely to be the causative mutation, which is as yet unknown. PMID:20345892
Adams, Temitope F; Wongchai, Chatchawal; Chaidee, Anchalee; Pfeiffer, Wolfgang
2016-01-01
Plant essential oils have been suggested as a promising alternative to the established mosquito repellent DEET (N,N-diethyl-meta-toluamide). Searching for an assay with generally available equipment, we designed a new audiovisual assay of repellent activity against mosquitoes "Singing in the Tube," testing single mosquitoes in Drosophila cultivation tubes. Statistics with regression analysis should compensate for limitations of simple hardware. The assay was established with female Culex pipiens mosquitoes in 60 experiments, 120-h audio recording, and 2580 estimations of the distance between mosquito sitting position and the chemical. Correlations between parameters of sitting position, flight activity pattern, and flight tone spectrum were analyzed. Regression analysis of psycho-acoustic data of audio files (dB[A]) used a squared and modified sinus function determining wing beat frequency WBF ± SD (357 ± 47 Hz). Application of logistic regression defined the repelling velocity constant. The repelling velocity constant showed a decreasing order of efficiency of plant essential oils: rosemary (Rosmarinus officinalis), eucalyptus (Eucalyptus globulus), lavender (Lavandula angustifolia), citronella (Cymbopogon nardus), tea tree (Melaleuca alternifolia), clove (Syzygium aromaticum), lemon (Citrus limon), patchouli (Pogostemon cablin), DEET, cedar wood (Cedrus atlantica). In conclusion, we suggest (1) disease vector control (e.g., impregnation of bed nets) by eight plant essential oils with repelling velocity superior to DEET, (2) simple mosquito repellency testing in Drosophila cultivation tubes, (3) automated approaches and room surveillance by generally available audio equipment (dB[A]: ISO standard 226), and (4) quantification of repellent activity by parameters of the audiovisual assay defined by correlation and regression analyses.
NASA Astrophysics Data System (ADS)
Gaydecki, Patrick; Fernandes, Bosco
2003-11-01
A fast digital signal processing (DSP) system is described that can perform real-time emulation of a wide variety of linear audio-bandwidth systems and networks, such as reverberant spaces, musical instrument bodies and very high order filter networks. The hardware design is based upon a Motorola DSP56309 operating at 110 million multiplication-accumulations per second and a dual-channel 24 bit codec with a maximum sampling frequency of 192 kHz. High level software has been developed to express complex vector frequency responses as both infinite impulse response (IIR) and finite impulse response (FIR) coefficients, in a form suitable for real-time convolution by the firmware installed in the DSP system memory. An algorithm has also been devised to express IIR filters as equivalent FIR structures, thereby obviating the potential instabilities associated with recursive equations and negating the traditional deficiencies of FIR filters respecting equivalent analogue designs. The speed and dynamic range of the system is such that, when sampling at 48 kHz, the frequency response can be specified to a spectral precision of 22 Hz when sampling at 10 kHz, this resolution increases to 0.9 Hz. Moreover, it is also possible to control the phase of any frequency band with a theoretical precision of 10-5 degrees in all cases. The system has been applied in the study of analogue filter networks, real-time Hilbert transformation, phase-shift systems and musical instrument body emulation, where it is providing valuable new insights into the understanding of psychoacoustic mechanisms.
Control of multi-joint arm movements for the manipulation of touch in keystroke by expert pianists
2010-01-01
Background Production of a variety of finger-key touches in the piano is essential for expressive musical performance. However, it remains unknown how expert pianists control multi-joint finger and arm movements for manipulating the touch. The present study investigated differences in kinematics and kinetics of the upper-limb movements while expert pianists were depressing a key with two different touches: pressed and struck. The former starts key-depression with the finger-tip contacting the key, whereas the latter involves preparatory arm-lift before striking the key. To determine the effect of individual muscular torque (MUS) as well as non-muscular torques on joint acceleration, we performed a series of inverse and forward dynamics computations. Results The pressed touch showed smaller elbow extension velocity, and larger shoulder and finger flexion velocities during key-depression compared with the struck touch. The former touch also showed smaller elbow extension acceleration directly attributed to the shoulder MUS. In contrast, the shoulder flexion acceleration induced by elbow and wrist MUS was greater for the pressed touch than the struck touch. Towards the goal of producing the target finger-key contact dynamics, the pressed and struck touches effectively took advantage of the distal-to-proximal and proximal-to-distal inter-segmental dynamics, respectively. Furthermore, a psychoacoustic experiment confirmed that a tone elicited by the pressed touch was perceived softer than that by the struck touch. Conclusion The present findings suggest that manipulation of tone timbre depends on control of inter-segmental dynamics in piano keystroke. PMID:20630085
Kollmeier, Birger; Schädler, Marc René; Warzybok, Anna; Meyer, Bernd T; Brand, Thomas
2016-09-07
To characterize the individual patient's hearing impairment as obtained with the matrix sentence recognition test, a simulation Framework for Auditory Discrimination Experiments (FADE) is extended here using the Attenuation and Distortion (A+D) approach by Plomp as a blueprint for setting the individual processing parameters. FADE has been shown to predict the outcome of both speech recognition tests and psychoacoustic experiments based on simulations using an automatic speech recognition system requiring only few assumptions. It builds on the closed-set matrix sentence recognition test which is advantageous for testing individual speech recognition in a way comparable across languages. Individual predictions of speech recognition thresholds in stationary and in fluctuating noise were derived using the audiogram and an estimate of the internal level uncertainty for modeling the individual Plomp curves fitted to the data with the Attenuation (A-) and Distortion (D-) parameters of the Plomp approach. The "typical" audiogram shapes from Bisgaard et al with or without a "typical" level uncertainty and the individual data were used for individual predictions. As a result, the individualization of the level uncertainty was found to be more important than the exact shape of the individual audiogram to accurately model the outcome of the German Matrix test in stationary or fluctuating noise for listeners with hearing impairment. The prediction accuracy of the individualized approach also outperforms the (modified) Speech Intelligibility Index approach which is based on the individual threshold data only. © The Author(s) 2016.
Rationale for the tinnitus retraining therapy trial
Formby, Craig; Scherer, Roberta
2013-01-01
The Tinnitus Retraining Therapy Trial (TRTT) is a National Institutes of Health-sponsored, multi-centered, placebo-controlled, randomized trial evaluating the efficacy of tinnitus retraining therapy (TRT) and its component parts, directive counseling and sound therapy, as treatments for subjective debilitating tinnitus in the military. The TRTT will enroll 228 individuals at an allocation ratio of 1:1:1 to: (1) directive counseling and sound therapy using conventional sound generators; (2) directive counseling and placebo sound generators; or (3) standard of care as administered in the military. Study centers include a Study Chair’s Office, a Data Coordinating Center, and six Military Clinical Centers with treatment and data collection standardized across all clinics. The primary outcome is change in Tinnitus Questionnaire (TQ) score assessed longitudinally at 3, 6, 12, and 18-month follow-up visits. Secondary outcomes include: Change in TQ sub-scales, Tinnitus Handicap Inventory, Tinnitus Functional Index, and TRT interview visual analog scale; audiometric and psychoacoustic measures; and change in quality of life. The TRTT will evaluate TRT efficacy by comparing TRT (directive counseling and conventional sound generators) with standard of care; directive counseling by comparing directive counseling plus placebo sound generators versus standard of care; and sound therapy by comparing conventional versus placebo sound generators. We hypothesize that full TRT will be more efficacious than standard of care, directive counseling and placebo sound generators more efficacious than standard of care, and conventional more efficacious than placebo sound generators in habituating the tinnitus awareness, annoyance, and impact on the study participant’s life. PMID:23571304
Erb, Julia; Ludwig, Alexandra Annemarie; Kunke, Dunja; Fuchs, Michael; Obleser, Jonas
2018-04-24
Psychoacoustic tests assessed shortly after cochlear implantation are useful predictors of the rehabilitative speech outcome. While largely independent, both spectral and temporal resolution tests are important to provide an accurate prediction of speech recognition. However, rapid tests of temporal sensitivity are currently lacking. Here, we propose a simple amplitude modulation rate discrimination (AMRD) paradigm that is validated by predicting future speech recognition in adult cochlear implant (CI) patients. In 34 newly implanted patients, we used an adaptive AMRD paradigm, where broadband noise was modulated at the speech-relevant rate of ~4 Hz. In a longitudinal study, speech recognition in quiet was assessed using the closed-set Freiburger number test shortly after cochlear implantation (t0) as well as the open-set Freiburger monosyllabic word test 6 months later (t6). Both AMRD thresholds at t0 (r = -0.51) and speech recognition scores at t0 (r = 0.56) predicted speech recognition scores at t6. However, AMRD and speech recognition at t0 were uncorrelated, suggesting that those measures capture partially distinct perceptual abilities. A multiple regression model predicting 6-month speech recognition outcome with deafness duration and speech recognition at t0 improved from adjusted R = 0.30 to adjusted R = 0.44 when AMRD threshold was added as a predictor. These findings identify AMRD thresholds as a reliable, nonredundant predictor above and beyond established speech tests for CI outcome. This AMRD test could potentially be developed into a rapid clinical temporal-resolution test to be integrated into the postoperative test battery to improve the reliability of speech outcome prognosis.
Kastelein, Ronald A; Hoek, Lean; Wensveen, Paul J; Terhune, John M; de Jong, Christ A F
2010-02-01
The underwater hearing sensitivities of two 2-year-old female harbor seals were quantified in a pool built for acoustic research by using a behavioral psycho-acoustic technique. The animals were trained only to respond when they detected an acoustic signal ("go/no-go" response). Detection thresholds were obtained for pure tone signals (frequencies: 0.2-40 kHz; durations: 0.5-5000 ms, depending on the frequency; 59 frequency-duration combinations). Detection thresholds were quantified by varying the signal amplitude by the 1-up, 1-down staircase method, and were defined as the stimulus levels, resulting in a 50% detection rate. The hearing thresholds of the two seals were similar for all frequencies except for 40 kHz, for which the thresholds differed by, on average, 3.7 dB. There was an inverse relationship between the time constant (tau), derived from an exponential model of temporal integration, and the frequency [log(tau)=2.86-0.94 log(f);tau in ms and f in kHz]. Similarly, the thresholds increased when the pulse was shorter than approximately 780 cycles (independent of the frequency). For pulses shorter than the integration time, the thresholds increased by 9-16 dB per decade reduction in the duration or number of cycles in the pulse. The results of this study suggest that most published hearing thresholds
Leftward Lateralization of Auditory Cortex Underlies Holistic Sound Perception in Williams Syndrome
Bendszus, Martin; Schneider, Peter
2010-01-01
Background Individuals with the rare genetic disorder Williams-Beuren syndrome (WS) are known for their characteristic auditory phenotype including strong affinity to music and sounds. In this work we attempted to pinpoint a neural substrate for the characteristic musicality in WS individuals by studying the structure-function relationship of their auditory cortex. Since WS subjects had only minor musical training due to psychomotor constraints we hypothesized that any changes compared to the control group would reflect the contribution of genetic factors to auditory processing and musicality. Methodology/Principal Findings Using psychoacoustics, magnetoencephalography and magnetic resonance imaging, we show that WS individuals exhibit extreme and almost exclusive holistic sound perception, which stands in marked contrast to the even distribution of this trait in the general population. Functionally, this was reflected by increased amplitudes of left auditory evoked fields. On the structural level, volume of the left auditory cortex was 2.2-fold increased in WS subjects as compared to control subjects. Equivalent volumes of the auditory cortex have been previously reported for professional musicians. Conclusions/Significance There has been an ongoing debate in the neuroscience community as to whether increased gray matter of the auditory cortex in musicians is attributable to the amount of training or innate disposition. In this study musical education of WS subjects was negligible and control subjects were carefully matched for this parameter. Therefore our results not only unravel the neural substrate for this particular auditory phenotype, but in addition propose WS as a unique genetic model for training-independent auditory system properties. PMID:20808792
The Gap Detection Test: Can It Be Used to Diagnose Tinnitus?
Boyen, Kris; Başkent, Deniz
2015-01-01
Objectives: Animals with induced tinnitus showed difficulties in detecting silent gaps in sounds, suggesting that the tinnitus percept may be filling the gap. The main purpose of this study was to evaluate the applicability of this approach to detect tinnitus in human patients. The authors first hypothesized that gap detection would be impaired in patients with tinnitus, and second, that gap detection would be more impaired at frequencies close to the tinnitus frequency of the patient. Design: Twenty-two adults with bilateral tinnitus, 20 age-matched and hearing loss–matched subjects without tinnitus, and 10 young normal-hearing subjects participated in the study. To determine the characteristics of the tinnitus, subjects matched an external sound to their perceived tinnitus in pitch and loudness. To determine the minimum detectable gap, the gap threshold, an adaptive psychoacoustic test was performed three times by each subject. In this gap detection test, four different stimuli, with various frequencies and bandwidths, were presented at three intensity levels each. Results: Similar to previous reports of gap detection, increasing sensation level yielded shorter gap thresholds for all stimuli in all groups. Interestingly, the tinnitus group did not display elevated gap thresholds in any of the four stimuli. Moreover, visual inspection of the data revealed no relation between gap detection performance and perceived tinnitus pitch. Conclusions: These findings show that tinnitus in humans has no effect on the ability to detect gaps in auditory stimuli. Thus, the testing procedure in its present form is not suitable for clinical detection of tinnitus in humans. PMID:25822647
Steel, Morrison M; Papsin, Blake C; Gordon, Karen A
2015-01-01
Bilateral cochlear implants aim to provide hearing to both ears for children who are deaf and promote binaural/spatial hearing. Benefits are limited by mismatched devices and unilaterally-driven development which could compromise the normal integration of left and right ear input. We thus asked whether children hear a fused image (ie. 1 vs 2 sounds) from their bilateral implants and if this "binaural fusion" reduces listening effort. Binaural fusion was assessed by asking 25 deaf children with cochlear implants and 24 peers with normal hearing whether they heard one or two sounds when listening to bilaterally presented acoustic click-trains/electric pulses (250 Hz trains of 36 ms presented at 1 Hz). Reaction times and pupillary changes were recorded simultaneously to measure listening effort. Bilaterally implanted children heard one image of bilateral input less frequently than normal hearing peers, particularly when intensity levels on each side were balanced. Binaural fusion declined as brainstem asymmetries increased and age at implantation decreased. Children implanted later had access to acoustic input prior to implantation due to progressive deterioration of hearing. Increases in both pupil diameter and reaction time occurred as perception of binaural fusion decreased. Results indicate that, without binaural level cues, children have difficulty fusing input from their bilateral implants to perceive one sound which costs them increased listening effort. Brainstem asymmetries exacerbate this issue. By contrast, later implantation, reflecting longer access to bilateral acoustic hearing, may have supported development of auditory pathways underlying binaural fusion. Improved integration of bilateral cochlear implant signals for children is required to improve their binaural hearing.
Marques do Carmo, Diego; Costa, Márcio Holsbach
2018-04-01
This work presents an online approximation method for the multichannel Wiener filter (MWF) noise reduction technique with preservation of the noise interaural level difference (ILD) for binaural hearing-aids. The steepest descent method is applied to a previously proposed MWF-ILD cost function to both approximate the optimal linear estimator of the desired speech and keep the subjective perception of the original acoustic scenario. The computational cost of the resulting algorithm is estimated in terms of multiply and accumulate operations, whose number can be controlled by setting the number of iterations at each time frame. Simulation results for the particular case of one speech and one-directional noise source show that the proposed method increases the signal-to-noise ratio SNR of the originally acquired speech by up to 16.9 dB in the assessed scenarios. As compared to the online implementation of the conventional MWF technique, the proposed technique provides a reduction of up to 7 dB in the noise ILD error at the price of a reduction of up 3 dB in the output SNR. Subjective experiments with volunteers complement these objective measures with psychoacoustic results, which corroborate the expected spatial preservation of the original acoustic scenario. The proposed method allows practical online implementation of the MWF-ILD noise reduction technique under constrained computational resources. Predicted SNR improvements from 12 dB to 16.9 dB can be obtained in application-specific integrated circuits for hearing-aids and state-of-the-art digital signal processors. Copyright © 2018 Elsevier Ltd. All rights reserved.
Meeuws, Matthias; Pascoal, David; Bermejo, Iñigo; Artaso, Miguel; De Ceulaer, Geert; Govaerts, Paul J
2017-07-01
The software application FOX ('Fitting to Outcome eXpert') is an intelligent agent to assist in the programing of cochlear implant (CI) processors. The current version utilizes a mixture of deterministic and probabilistic logic which is able to improve over time through a learning effect. This study aimed at assessing whether this learning capacity yields measurable improvements in speech understanding. A retrospective study was performed on 25 consecutive CI recipients with a median CI use experience of 10 years who came for their annual CI follow-up fitting session. All subjects were assessed by means of speech audiometry with open set monosyllables at 40, 55, 70, and 85 dB SPL in quiet with their home MAP. Other psychoacoustic tests were executed depending on the audiologist's clinical judgment. The home MAP and the corresponding test results were entered into FOX. If FOX suggested to make MAP changes, they were implemented and another speech audiometry was performed with the new MAP. FOX suggested MAP changes in 21 subjects (84%). The within-subject comparison showed a significant median improvement of 10, 3, 1, and 7% at 40, 55, 70, and 85 dB SPL, respectively. All but two subjects showed an instantaneous improvement in their mean speech audiometric score. Persons with long-term CI use, who received a FOX-assisted CI fitting at least 6 months ago, display improved speech understanding after MAP modifications, as recommended by the current version of FOX. This can be explained only by intrinsic improvements in FOX's algorithms, as they have resulted from learning. This learning is an inherent feature of artificial intelligence and it may yield measurable benefit in speech understanding even in long-term CI recipients.
Klotz, Sebastian
2008-09-01
The study of acoustics, harmonics and of music has been providing scientific models since Greek Antiquity. Since the early modern ages, two separate cultures began to emerge out of the study of music: a technical acoustics and an aesthetically and philosophically inspired musical criticism. In the writings of Johann Friedrich Herbart (1811) a scientific approach to musical aesthetics and to music perception is taking shape that reinstalls the listening process as a highly complex and logical phenomenon. By opening music for a scientific psychological investigation, Herbart pioneered the physiologically and acoustically grounded seminal work by Hermann von Helmholtz On the sensations of tone (1863) which the author considered a prerequisite for musical aesthetics and music theory. Helmholtz in turn inspired the philosopher and psychologist Carl Stumpf to further investigate musical perception (beginning in 1883). To Stumpf, it provided a paradigm for experimental psychology as mental functions and phenomena could be studied in detail. These functions and phenomena are the actual objects of scientific study in Stumpf's inductive and descriptive psychology. Combining insights from statistics, ethnology, anthropology, psychoacoustics and the cultural history of mankind, Stumpf and his team developed a new blend of science which absorbs styles of reasoning, analytical procedures and academic convictions from natural history, the natural sciences and the humanities but at the same time identifies shortcomings of these approaches that fail to grasp the complexities of psychic functions. Despite their reliance on the quasi-objective phonograph and despite their commitment to objectivity, precision and measurement, mental phenomena relating to tonal perception and to music provided too complex a challenge to be easily articulated and shared by the scientific community after 1900. The essay illustrates these tensions against the background of a history of objectivity.
Fröhlich, Felix; Ernst, Arne; Strübing, Ira; Basta, Dietmar; Gröschel, Moritz
2017-12-01
A correlation between noise-induced apoptosis and cell loss has previously been shown after a single noise exposure in the cochlear nucleus, inferior colliculus, medial geniculate body (MGB) and primary auditory cortex (AI). However, repeated noise exposure is the most common situation in humans and a major risk factor for the induction of noise-induced hearing loss (NIHL). The present investigation measured cell death pathways using terminal deoxynucleotidyl transferase dUTP nick end labeling (TUNEL) in the dorsal, medial and ventral MGB (dMGB, mMGB and vMGB) and six layers of the AI (AI-1 to AI-6) in mice (NMRI strain) after a second noise exposure (double-exposure group). Therefore, a single noise exposure group has been investigated 7 (7-day-group-single) or 14 days (14-day-group-single) after noise exposure (3 h, 5-20 kHz, 115 dB SPL peak-to-peak). The double-exposure group received the same noise trauma for a second time 7 days after the initial exposure and was either TUNEL-stained immediately (7-day-group-double) or 1 week later (14-day-group-double) and data were compared to the corresponding single-trauma group as well as to an unexposed control group. It was shown that TUNEL increased immediately after the second noise exposure in AI-3 and stayed upregulated in the 14-day-group-double. A significant increase in TUNEL was also seen in the 14-day-group-double in vMGB, mMGB and AI-1. The present results show for the first time the influence of a repeated noise trauma on cell death mechanisms in thalamic and cortical structures and might contribute to the understanding of pathophysiological findings and psychoacoustic phenomena accompanying NIHL.
NASA Astrophysics Data System (ADS)
Charbonneau, Jeremy
As the perceived quality of a product is becoming more important in the manufacturing industry, more emphasis is being placed on accurately predicting the sound quality of everyday objects. This study was undertaken to improve upon current prediction techniques with regard to the psychoacoustic descriptor of loudness and an improved binaural summation technique. The feasibility of this project was first investigated through a loudness matching experiment involving thirty-one subjects and pure tones of constant sound pressure level. A dependence of binaural summation on frequency was observed which had previously not been a subject of investigation in the reviewed literature. A follow-up investigation was carried out with forty-eight volunteers and pure tones of constant sensation level. Contrary to existing theories in literature the resulting loudness matches revealed an amplitude versus frequency relationship which confirmed the perceived increase in loudness when a signal was presented to both ears simultaneously as opposed to one ear alone. The resulting trend strongly indicated that the higher the frequency of the presented signal, the greater the increase in observed binaural summation. The results from each investigation were summarized into a single binaural summation algorithm and inserted into an improved time-varying loudness model. Using experimental techniques, it was demonstrated that the updated binaural summation algorithm was a considerable improvement over the state of the art approach for predicting the perceived binaural loudness. The improved function retained the ease of use from the original model while additionally providing accurate estimates of diotic listening conditions from monaural WAV files. It was clearly demonstrated using a validation jury test that the revised time-varying loudness model was a significant improvement over the previously standardized approach.
Shetty, Hemanth Narayan; Koonoor, Vishal
2016-11-01
Past research has reported that children with repeated occurrences of otitis media at an early age have a negative impact on speech perception at a later age. The present study necessitates documenting the temporal and spectral processing on speech perception in noise from normal and atypical groups. The present study evaluated the relation between speech perception in noise and temporal; and spectral processing abilities in children with normal and atypical groups. The study included two experiments. In the first experiment, temporal resolution and frequency discrimination of listeners with normal group and three subgroups of atypical groups (had a history of OM) a) less than four episodes b) four to nine episodes and c) More than nine episodes during their chronological age of 6 months to 2 years) were evaluated using measures of temporal modulation transfer function and frequency discrimination test. In the second experiment, SNR 50 was evaluated on each group of study participants. All participants had normal hearing and middle ear status during the course of testing. Demonstrated that children with atypical group had significantly poorer modulation detection threshold, peak sensitivity and bandwidth; and frequency discrimination to each F0 than normal hearing listeners. Furthermore, there was a significant correlation seen between measures of temporal resolution; frequency discrimination and speech perception in noise. It infers atypical groups have significant impairment in extracting envelope as well as fine structure cues from the signal. The results supported the idea that episodes of OM before 2 years of agecan produce periods of sensory deprivation that alters the temporal and spectral skills which in turn has negative consequences on speech perception in noise. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Nikolsky, Aleksey
2015-01-01
This paper reveals the way in which musical pitch works as a peculiar form of cognition that reflects upon the organization of the surrounding world as perceived by majority of music users within a socio-cultural formation. The evidence from music theory, ethnography, archeology, organology, anthropology, psychoacoustics, and evolutionary biology is plotted against experimental evidence. Much of the methodology for this investigation comes from studies conducted within the territory of the former USSR. To date, this methodology has remained solely confined to Russian speaking scholars. A brief overview of pitch-set theory demonstrates the need to distinguish between vertical and horizontal harmony, laying out the framework for virtual music space that operates according to the perceptual laws of tonal gravity. Brought to life by bifurcation of music and speech, tonal gravity passed through eleven discrete stages of development until the onset of tonality in the seventeenth century. Each stage presents its own method of integration of separate musical tones into an auditory-cognitive unity. The theory of "melodic intonation" is set forth as a counterpart to harmonic theory of chords. Notions of tonality, modality, key, diatonicity, chromaticism, alteration, and modulation are defined in terms of their perception, and categorized according to the way in which they have developed historically. Tonal organization in music, and perspective organization in fine arts are explained as products of the same underlying mental process. Music seems to act as a unique medium of symbolic representation of reality through the concept of pitch. Tonal organization of pitch reflects the culture of thinking, adopted as a standard within a community of music users. Tonal organization might be a naturally formed system of optimizing individual perception of reality within a social group and its immediate environment, setting conventional standards of intellectual and emotional intelligence.
Self-reports about tinnitus and about cochlear implants.
Noble, W
2000-08-01
Analyze literature on self-report outcomes in two areas of audiological rehabilitation: 1) tinnitus and 2) cochlear implant hearing aids. 1) Tinnitus: survey of features in the development of self-report approaches and of formal scales used in assessment of tinnitus disability and handicaps. 2) Cochlear implants: summary of the literature using self-report approaches to cochlear implant experience that indicates points of theoretical significance. 1) Major features of tinnitus are: a) disabilities such as interference with and distortion of normal auditory perception; b) handicaps such as emotional distress, interference with sleep, and with personal and social life. Nonauditory factors-chronic depression, high self-focused attention-mediate the degree of experienced tinnitus handicap. 2) People with prelingual loss of hearing report that a cochlear implant primarily enables improved detection and discrimination of environmental sound; those with postlingual loss find that an implant in addition provides improved speech recognition. 1) Coping with tinnitus is influenced by the personal resources that can be brought to bear on the experience, highlighting a general point that any rehabilitation outcome is not only a matter of acoustical solutions. By the same token, tinnitus can be easier to cope with if its "psychoacoustic presence" can be diminished by some form of masking. 2) Cochlear implants fitted in childhood that do not provide meaningful input signals in real-world settings may be rejected in adolescence. 3) "Hearing," as a capacity, does not have a fixed worth. Different circumstances mean it will be taken as desirable or as delivering torment (extreme tinnitus, e.g.). Its value will also vary depending on the extent of a person's access to spoken language (aiding in very early childhood, e.g.).
Verhey, Jesko L; Epp, Bastian; Stasiak, Arkadiusz; Winter, Ian M
2013-01-01
A common characteristic of natural sounds is that the level fluctuations in different frequency regions are coherent. The ability of the auditory system to use this comodulation is shown when a sinusoidal signal is masked by a masker centred at the signal frequency (on-frequency masker, OFM) and one or more off-frequency components, commonly referred to as flanking bands (FBs). In general, the threshold of the signal masked by comodulated masker components is lower than when masked by masker components with uncorrelated envelopes or in the presence of the OFM only. This effect is commonly referred to as comodulation masking release (CMR). The present study investigates if CMR is also observed for a sinusoidal signal embedded in the OFM when the centre frequencies of the FBs are swept over time with a sweep rate of one octave per second. Both a common change of different frequencies and comodulation could serve as cues to indicate which of the stimulus components originate from one source. If the common fate of frequency components is the stronger binding cue, the sweeping FBs and the OFM with a fixed centre frequency should no longer form one auditory object and the CMR should be abolished. However, psychoacoustical results with normal-hearing listeners show that a CMR is also observed with sweeping components. The results are consistent with the hypothesis of wideband inhibition as the underlying physiological mechanism, as the CMR should only depend on the spectral position of the flanking bands relative to the inhibitory areas (as seen in physiological recordings using stationary flanking bands). Preliminary physiological results in the cochlear nucleus of the Guinea pig show that a correlate of CMR can also be found at this level of the auditory pathway with sweeping flanking bands.
NASA Astrophysics Data System (ADS)
Suyatno; Tjokronegoro, H. A.; Merthayasa, I. G. N.; Supanggah, R.
2016-11-01
This paper presents results of an investigation of room acoustic parameters those are appropriated particularly to perform a Javanese gamelan. The acoustic parameters were obtained by analysing simulated sounds of performance. Those simulated sounds were obtained by sound convolution technique of a dry sound signal of Javanese gamelan performance with impulse responses rooms whose appropriated reverberation time. The reverberation time were varied between 1.0s to 1.8s, those belong to the characteristic of Pendopo Mangkunegaran Surakarta. In this case, Pendopo Mangkunegaran is assumed as one of the most suitable concert halls for Javanese gamelan performance. To obtain the acoustic parameters, we used a psycho-acoustics measurement based on paired comparison test that having different of acoustic parameters to determine the most comfortable one to majority of respondents. The respondents who have participated in this research composed of a group of professional musicians of Javanese gamelan and groups of audience who are not musician, nevertheless part of them were familiar with Javanese gamelan music. The comparison test gave results and showed majority of respondents of group of musicians had a notion sound reverberation time of 1.2s was most comfortable. This corresponds to +6.2dB, clarity and 74% definition. It means the appropriate acoustic condition allows musicians to recognize and distinguish clearly sound of each instrument being played. Meanwhile, group of audience had a notion reverberation time in a range of 1.2s - 1.6s was most comfortable. This range of reverberation time corresponds to +4dB to +6.2dB of clarity, and 66% to 74% of definition.
Steel, Morrison M.; Papsin, Blake C.; Gordon, Karen A.
2015-01-01
Bilateral cochlear implants aim to provide hearing to both ears for children who are deaf and promote binaural/spatial hearing. Benefits are limited by mismatched devices and unilaterally-driven development which could compromise the normal integration of left and right ear input. We thus asked whether children hear a fused image (ie. 1 vs 2 sounds) from their bilateral implants and if this “binaural fusion” reduces listening effort. Binaural fusion was assessed by asking 25 deaf children with cochlear implants and 24 peers with normal hearing whether they heard one or two sounds when listening to bilaterally presented acoustic click-trains/electric pulses (250 Hz trains of 36 ms presented at 1 Hz). Reaction times and pupillary changes were recorded simultaneously to measure listening effort. Bilaterally implanted children heard one image of bilateral input less frequently than normal hearing peers, particularly when intensity levels on each side were balanced. Binaural fusion declined as brainstem asymmetries increased and age at implantation decreased. Children implanted later had access to acoustic input prior to implantation due to progressive deterioration of hearing. Increases in both pupil diameter and reaction time occurred as perception of binaural fusion decreased. Results indicate that, without binaural level cues, children have difficulty fusing input from their bilateral implants to perceive one sound which costs them increased listening effort. Brainstem asymmetries exacerbate this issue. By contrast, later implantation, reflecting longer access to bilateral acoustic hearing, may have supported development of auditory pathways underlying binaural fusion. Improved integration of bilateral cochlear implant signals for children is required to improve their binaural hearing. PMID:25668423
A New Test of Attention in Listening (TAIL) Predicts Auditory Performance
Zhang, Yu-Xuan; Barry, Johanna G.; Moore, David R.; Amitay, Sygal
2012-01-01
Attention modulates auditory perception, but there are currently no simple tests that specifically quantify this modulation. To fill the gap, we developed a new, easy-to-use test of attention in listening (TAIL) based on reaction time. On each trial, two clearly audible tones were presented sequentially, either at the same or different ears. The frequency of the tones was also either the same or different (by at least two critical bands). When the task required same/different frequency judgments, presentation at the same ear significantly speeded responses and reduced errors. A same/different ear (location) judgment was likewise facilitated by keeping tone frequency constant. Perception was thus influenced by involuntary orienting of attention along the task-irrelevant dimension. When information in the two stimulus dimensions were congruent (same-frequency same-ear, or different-frequency different-ear), response was faster and more accurate than when they were incongruent (same-frequency different-ear, or different-frequency same-ear), suggesting the involvement of executive control to resolve conflicts. In total, the TAIL yielded five independent outcome measures: (1) baseline reaction time, indicating information processing efficiency, (2) involuntary orienting of attention to frequency and (3) location, and (4) conflict resolution for frequency and (5) location. Processing efficiency and conflict resolution accounted for up to 45% of individual variances in the low- and high-threshold variants of three psychoacoustic tasks assessing temporal and spectral processing. Involuntary orientation of attention to the irrelevant dimension did not correlate with perceptual performance on these tasks. Given that TAIL measures are unlikely to be limited by perceptual sensitivity, we suggest that the correlations reflect modulation of perceptual performance by attention. The TAIL thus has the power to identify and separate contributions of different components of attention to auditory perception. PMID:23300934
Humes, Larry E.; Kidd, Gary R.; Lentz, Jennifer J.
2013-01-01
This study was designed to address individual differences in aided speech understanding among a relatively large group of older adults. The group of older adults consisted of 98 adults (50 female and 48 male) ranging in age from 60 to 86 (mean = 69.2). Hearing loss was typical for this age group and about 90% had not worn hearing aids. All subjects completed a battery of tests, including cognitive (6 measures), psychophysical (17 measures), and speech-understanding (9 measures), as well as the Speech, Spatial, and Qualities of Hearing (SSQ) self-report scale. Most of the speech-understanding measures made use of competing speech and the non-speech psychophysical measures were designed to tap phenomena thought to be relevant for the perception of speech in competing speech (e.g., stream segregation, modulation-detection interference). All measures of speech understanding were administered with spectral shaping applied to the speech stimuli to fully restore audibility through at least 4000 Hz. The measures used were demonstrated to be reliable in older adults and, when compared to a reference group of 28 young normal-hearing adults, age-group differences were observed on many of the measures. Principal-components factor analysis was applied successfully to reduce the number of independent and dependent (speech understanding) measures for a multiple-regression analysis. Doing so yielded one global cognitive-processing factor and five non-speech psychoacoustic factors (hearing loss, dichotic signal detection, multi-burst masking, stream segregation, and modulation detection) as potential predictors. To this set of six potential predictor variables were added subject age, Environmental Sound Identification (ESI), and performance on the text-recognition-threshold (TRT) task (a visual analog of interrupted speech recognition). These variables were used to successfully predict one global aided speech-understanding factor, accounting for about 60% of the variance. PMID:24098273
Auditory Time-Frequency Masking for Spectrally and Temporally Maximally-Compact Stimuli
Laback, Bernhard; Savel, Sophie; Ystad, Sølvi; Balazs, Peter; Meunier, Sabine; Kronland-Martinet, Richard
2016-01-01
Many audio applications perform perception-based time-frequency (TF) analysis by decomposing sounds into a set of functions with good TF localization (i.e. with a small essential support in the TF domain) using TF transforms and applying psychoacoustic models of auditory masking to the transform coefficients. To accurately predict masking interactions between coefficients, the TF properties of the model should match those of the transform. This involves having masking data for stimuli with good TF localization. However, little is known about TF masking for mathematically well-localized signals. Most existing masking studies used stimuli that are broad in time and/or frequency and few studies involved TF conditions. Consequently, the present study had two goals. The first was to collect TF masking data for well-localized stimuli in humans. Masker and target were 10-ms Gaussian-shaped sinusoids with a bandwidth of approximately one critical band. The overall pattern of results is qualitatively similar to existing data for long maskers. To facilitate implementation in audio processing algorithms, a dataset provides the measured TF masking function. The second goal was to assess the potential effect of auditory efferents on TF masking using a modeling approach. The temporal window model of masking was used to predict present and existing data in two configurations: (1) with standard model parameters (i.e. without efferents), (2) with cochlear gain reduction to simulate the activation of efferents. The ability of the model to predict the present data was quite good with the standard configuration but highly degraded with gain reduction. Conversely, the ability of the model to predict existing data for long maskers was better with than without gain reduction. Overall, the model predictions suggest that TF masking can be affected by efferent (or other) effects that reduce cochlear gain. Such effects were avoided in the experiment of this study by using maximally-compact stimuli. PMID:27875575
Auditory Time-Frequency Masking for Spectrally and Temporally Maximally-Compact Stimuli.
Necciari, Thibaud; Laback, Bernhard; Savel, Sophie; Ystad, Sølvi; Balazs, Peter; Meunier, Sabine; Kronland-Martinet, Richard
2016-01-01
Many audio applications perform perception-based time-frequency (TF) analysis by decomposing sounds into a set of functions with good TF localization (i.e. with a small essential support in the TF domain) using TF transforms and applying psychoacoustic models of auditory masking to the transform coefficients. To accurately predict masking interactions between coefficients, the TF properties of the model should match those of the transform. This involves having masking data for stimuli with good TF localization. However, little is known about TF masking for mathematically well-localized signals. Most existing masking studies used stimuli that are broad in time and/or frequency and few studies involved TF conditions. Consequently, the present study had two goals. The first was to collect TF masking data for well-localized stimuli in humans. Masker and target were 10-ms Gaussian-shaped sinusoids with a bandwidth of approximately one critical band. The overall pattern of results is qualitatively similar to existing data for long maskers. To facilitate implementation in audio processing algorithms, a dataset provides the measured TF masking function. The second goal was to assess the potential effect of auditory efferents on TF masking using a modeling approach. The temporal window model of masking was used to predict present and existing data in two configurations: (1) with standard model parameters (i.e. without efferents), (2) with cochlear gain reduction to simulate the activation of efferents. The ability of the model to predict the present data was quite good with the standard configuration but highly degraded with gain reduction. Conversely, the ability of the model to predict existing data for long maskers was better with than without gain reduction. Overall, the model predictions suggest that TF masking can be affected by efferent (or other) effects that reduce cochlear gain. Such effects were avoided in the experiment of this study by using maximally-compact stimuli.
Lindblad, Ann-Cathrine; Rosenhall, Ulf; Olofsson, Åke; Hagerman, Björn
2014-01-01
The aim of the investigation was to study if dysfunctions associated to the cochlea or its regulatory system can be found, and possibly explain hearing problems in subjects with normal or near-normal audiograms. The design was a prospective study of subjects recruited from the general population. The included subjects were persons with auditory problems who had normal, or near-normal, pure tone hearing thresholds, who could be included in one of three subgroups: teachers, Education; people working with music, Music; and people with moderate or negligible noise exposure, Other. A fourth group included people with poorer pure tone hearing thresholds and a history of severe occupational noise, Industry. Ntotal = 193. The following hearing tests were used: - pure tone audiometry with Békésy technique, - transient evoked otoacoustic emissions and distortion product otoacoustic emissions, without and with contralateral noise; - psychoacoustical modulation transfer function, - forward masking, - speech recognition in noise, - tinnitus matching. A questionnaire about occupations, noise exposure, stress/anxiety, muscular problems, medication, and heredity, was addressed to the participants. Forward masking results were significantly worse for Education and Industry than for the other groups, possibly associated to the inner hair cell area. Forward masking results were significantly correlated to louder matched tinnitus. For many subjects speech recognition in noise, left ear, did not increase in a normal way when the listening level was increased. Subjects hypersensitive to loud sound had significantly better speech recognition in noise at the lower test level than subjects not hypersensitive. Self-reported stress/anxiety was similar for all groups. In conclusion, hearing dysfunctions were found in subjects with tinnitus and other auditory problems, combined with normal or near-normal pure tone thresholds. The teachers, mostly regarded as a group exposed to noise below risk levels, had dysfunctions almost identical to those of the more exposed Industry group.
Johannesen, Peter T.; Pérez-González, Patricia; Lopez-Poveda, Enrique A.
2014-01-01
Identifying the multiple contributors to the audiometric loss of a hearing impaired (HI) listener at a particular frequency is becoming gradually more useful as new treatments are developed. Here, we infer the contribution of inner (IHC) and outer hair cell (OHC) dysfunction to the total audiometric loss in a sample of 68 hearing aid candidates with mild-to-severe sensorineural hearing loss, and for test frequencies of 0.5, 1, 2, 4, and 6 kHz. It was assumed that the audiometric loss (HLTOTAL) at each test frequency was due to a combination of cochlear gain loss, or OHC dysfunction (HLOHC), and inefficient IHC processes (HLIHC), all of them in decibels. HLOHC and HLIHC were estimated from cochlear I/O curves inferred psychoacoustically using the temporal masking curve (TMC) method. 325 I/O curves were measured and 59% of them showed a compression threshold (CT). The analysis of these I/O curves suggests that (1) HLOHC and HLIHC account on average for 60–70 and 30–40% of HLTOTAL, respectively; (2) these percentages are roughly constant across frequencies; (3) across-listener variability is large; (4) residual cochlear gain is negatively correlated with hearing loss while residual compression is not correlated with hearing loss. Altogether, the present results support the conclusions from earlier studies and extend them to a wider range of test frequencies and hearing-loss ranges. Twenty-four percent of I/O curves were linear and suggested total cochlear gain loss. The number of linear I/O curves increased gradually with increasing frequency. The remaining 17% I/O curves suggested audiometric losses due mostly to IHC dysfunction and were more frequent at low (≤1 kHz) than at high frequencies. It is argued that in a majority of listeners, hearing loss is due to a common mechanism that concomitantly alters IHC and OHC function and that IHC processes may be more labile in the apex than in the base. PMID:25100940
Aucouturier, Jean-Julien; Canonne, Clément
2017-04-01
A recently emerging view in music cognition holds that music is not only social and participatory in its production, but also in its perception, i.e. that music is in fact perceived as the sonic trace of social relations between a group of real or virtual agents. While this view appears compatible with a number of intriguing music cognitive phenomena, such as the links between beat entrainment and prosocial behaviour or between strong musical emotions and empathy, direct evidence is lacking that listeners are at all able to use the acoustic features of a musical interaction to infer the affiliatory or controlling nature of an underlying social intention. We created a novel experimental situation in which we asked expert music improvisers to communicate 5 types of non-musical social intentions, such as being domineering, disdainful or conciliatory, to one another solely using musical interaction. Using a combination of decoding studies, computational and psychoacoustical analyses, we show that both musically-trained and non musically-trained listeners can recognize relational intentions encoded in music, and that this social cognitive ability relies, to a sizeable extent, on the information processing of acoustic cues of temporal and harmonic coordination that are not present in any one of the musicians' channels, but emerge from the dynamics of their interaction. By manipulating these cues in two-channel audio recordings and testing their impact on the social judgements of non-musician observers, we finally establish a causal relationship between the affiliation dimension of social behaviour and musical harmonic coordination on the one hand, and between the control dimension and musical temporal coordination on the other hand. These results provide novel mechanistic insights not only into the social cognition of musical interactions, but also into that of non-verbal interactions as a whole. Copyright © 2017. Published by Elsevier B.V.
Rocha, Andressa Vital; Mondelli, Maria Fernanda Capoani Garcia
The relations between the tinnitus and the hearing loss are due to the sensory deprivation caused by hearing loss, since this is followed by the functional and structural alteration of the auditory system as a whole. The cochlear lesions are accompanied by a reduction in the activity of the cochlear nerve, and the neural activity keeps increased in mainly all the central auditory nervous system to compensate this deficit. This study aimed to verify the effectiveness of the sound generator (SG) associated with the counseling in the treatment of the tinnitus in individuals with and without hearing loss regarding the improvement of the nuisance through Tinnitus Handicap Inventory (THI) and Visual Analogue Scale (VAS). The sample consisted of 30 individuals of both genders divided into two groups: Group 1 (G1) was comprised of 15 individuals with tinnitus and normal hearing, adapted to SG; Group 2 (G2) was comprised of 15 individuals with complaints of hearing acuity and tinnitus, adapted with SG and an individual hearing aid device (HA). Both groups underwent the following procedures: anamnesis and history of complaint, high frequency audiometry (HFA), imitanciometry, acuphenometry with the survey of psychoacoustic pitch and loudness thresholds and application of the tools THI and VAS. All of them were adapted with HA and Siemens SG and participated in a session of counseling. The individuals were assessed in three situations: initial assessment (before the adaptation of the HA and SG), monitoring and final assessment (6 months after adaptation). The comparison of the tinnitus nuisance and handicap in the three stages of assessment showed a significant improvement for both groups. The use of the SG was similarly effective in the treatment of the tinnitus in individuals with and without hearing loss, causing an improvement of the nuisance and handicap. Copyright © 2016 Associação Brasileira de Otorrinolaringologia e Cirurgia Cérvico-Facial. Published by Elsevier Editora Ltda. All rights reserved.
Lindblad, Ann-Cathrine; Rosenhall, Ulf; Olofsson, Åke; Hagerman, Björn
2014-01-01
The aim of the investigation was to study if dysfunctions associated to the cochlea or its regulatory system can be found, and possibly explain hearing problems in subjects with normal or near-normal audiograms. The design was a prospective study of subjects recruited from the general population. The included subjects were persons with auditory problems who had normal, or near-normal, pure tone hearing thresholds, who could be included in one of three subgroups: teachers, Education; people working with music, Music; and people with moderate or negligible noise exposure, Other. A fourth group included people with poorer pure tone hearing thresholds and a history of severe occupational noise, Industry. Ntotal = 193. The following hearing tests were used: − pure tone audiometry with Békésy technique, − transient evoked otoacoustic emissions and distortion product otoacoustic emissions, without and with contralateral noise; − psychoacoustical modulation transfer function, − forward masking, − speech recognition in noise, − tinnitus matching. A questionnaire about occupations, noise exposure, stress/anxiety, muscular problems, medication, and heredity, was addressed to the participants. Forward masking results were significantly worse for Education and Industry than for the other groups, possibly associated to the inner hair cell area. Forward masking results were significantly correlated to louder matched tinnitus. For many subjects speech recognition in noise, left ear, did not increase in a normal way when the listening level was increased. Subjects hypersensitive to loud sound had significantly better speech recognition in noise at the lower test level than subjects not hypersensitive. Self-reported stress/anxiety was similar for all groups. In conclusion, hearing dysfunctions were found in subjects with tinnitus and other auditory problems, combined with normal or near-normal pure tone thresholds. The teachers, mostly regarded as a group exposed to noise below risk levels, had dysfunctions almost identical to those of the more exposed Industry group. PMID:24827149
Advances in edge-diffraction modeling for virtual-acoustic simulations
NASA Astrophysics Data System (ADS)
Calamia, Paul Thomas
In recent years there has been growing interest in modeling sound propagation in complex, three-dimensional (3D) virtual environments. With diverse applications for the military, the gaming industry, psychoacoustics researchers, architectural acousticians, and others, advances in computing power and 3D audio-rendering techniques have driven research and development aimed at closing the gap between the auralization and visualization of virtual spaces. To this end, this thesis focuses on improving the physical and perceptual realism of sound-field simulations in virtual environments through advances in edge-diffraction modeling. To model sound propagation in virtual environments, acoustical simulation tools commonly rely on geometrical-acoustics (GA) techniques that assume asymptotically high frequencies, large flat surfaces, and infinitely thin ray-like propagation paths. Such techniques can be augmented with diffraction modeling to compensate for the effect of surface size on the strength and directivity of a reflection, to allow for propagation around obstacles and into shadow zones, and to maintain soundfield continuity across reflection and shadow boundaries. Using a time-domain, line-integral formulation of the Biot-Tolstoy-Medwin (BTM) diffraction expression, this thesis explores various aspects of diffraction calculations for virtual-acoustic simulations. Specifically, we first analyze the periodic singularity of the BTM integrand and describe the relationship between the singularities and higher-order reflections within wedges with open angle less than 180°. Coupled with analytical approximations for the BTM expression, this analysis allows for accurate numerical computations and a continuous sound field in the vicinity of an arbitrary wedge geometry insonified by a point source. Second, we describe an edge-subdivision strategy that allows for fast diffraction calculations with low error relative to a numerically more accurate solution. Third, to address the considerable increase in propagation paths due to diffraction, we describe a simple procedure for identifying and culling insignificant diffraction components during a virtual-acoustic simulation. Finally, we present a novel method to find GA components using diffraction parameters that ensures continuity at reflection and shadow boundaries.
Long-term tinnitus suppression with linear octave frequency transposition hearing AIDS.
Peltier, Elisabeth; Peltier, Cedric; Tahar, Stephanie; Alliot-Lugaz, Evelyne; Cazals, Yves
2012-01-01
Over the last three years of hearing aid dispensing, it was observed that among 74 subjects fitted with a linear octave frequency transposition (LOFT) hearing aid, 60 reported partial or complete tinnitus suppression during day and night, an effect still lasting after several months or years of daily use. We report in more details on 38 subjects from whom we obtained quantified measures of tinnitus suppression through visual analog scaling and several additional psychoacoustic and audiometric measures. The long-term suppression seems independent of subject age, and of duration and subjective localization of tinnitus. A small but significant correlation was found with audiogram losses but not with high frequency loss slope. Long-term tinnitus suppression was observed for different etiologies, but with a low success rate for sudden deafness. It should be noted that a majority of subjects (23) had a history of noise exposure. Tinnitus suppression started after a few days of LOFT hearing aid use and reached a maximum after a few weeks of daily use. For nine subjects different amounts of frequency shifting were tried and found more or less successful for long-term tinnitus suppression, no correlation was found with tinnitus pitch. When the use of the LOFT hearing aid was stopped tinnitus reappeared within a day, and after re-using the LOFT aid it disappeared again within a day. For about one third of the 38 subjects a classical amplification or a non linear frequency compression aid was also tried, and no such tinnitus suppression was observed. Besides improvements in audiometric sensitivity to high frequencies and in speech discrimination scores, LOFT can be considered as a remarkable opportunity to suppress tinnitus over a long time scale. From a pathophysiological viewpoint these observations seem to fit with a possible re-attribution of activity to previously deprived cerebral areas corresponding to high frequency coding.
Nikolsky, Aleksey
2015-01-01
This paper reveals the way in which musical pitch works as a peculiar form of cognition that reflects upon the organization of the surrounding world as perceived by majority of music users within a socio-cultural formation. The evidence from music theory, ethnography, archeology, organology, anthropology, psychoacoustics, and evolutionary biology is plotted against experimental evidence. Much of the methodology for this investigation comes from studies conducted within the territory of the former USSR. To date, this methodology has remained solely confined to Russian speaking scholars. A brief overview of pitch-set theory demonstrates the need to distinguish between vertical and horizontal harmony, laying out the framework for virtual music space that operates according to the perceptual laws of tonal gravity. Brought to life by bifurcation of music and speech, tonal gravity passed through eleven discrete stages of development until the onset of tonality in the seventeenth century. Each stage presents its own method of integration of separate musical tones into an auditory-cognitive unity. The theory of “melodic intonation” is set forth as a counterpart to harmonic theory of chords. Notions of tonality, modality, key, diatonicity, chromaticism, alteration, and modulation are defined in terms of their perception, and categorized according to the way in which they have developed historically. Tonal organization in music, and perspective organization in fine arts are explained as products of the same underlying mental process. Music seems to act as a unique medium of symbolic representation of reality through the concept of pitch. Tonal organization of pitch reflects the culture of thinking, adopted as a standard within a community of music users. Tonal organization might be a naturally formed system of optimizing individual perception of reality within a social group and its immediate environment, setting conventional standards of intellectual and emotional intelligence. PMID:26528193
An integrated system for dynamic control of auditory perspective in a multichannel sound field
NASA Astrophysics Data System (ADS)
Corey, Jason Andrew
An integrated system providing dynamic control of sound source azimuth, distance and proximity to a room boundary within a simulated acoustic space is proposed for use in multichannel music and film sound production. The system has been investigated, implemented, and psychoacoustically tested within the ITU-R BS.775 recommended five-channel (3/2) loudspeaker layout. The work brings together physical and perceptual models of room simulation to allow dynamic placement of virtual sound sources at any location of a simulated space within the horizontal plane. The control system incorporates a number of modules including simulated room modes, "fuzzy" sources, and tracking early reflections, whose parameters are dynamically changed according to sound source location within the simulated space. The control functions of the basic elements, derived from theories of perception of a source in a real room, have been carefully tuned to provide efficient, effective, and intuitive control of a sound source's perceived location. Seven formal listening tests were conducted to evaluate the effectiveness of the algorithm design choices. The tests evaluated: (1) loudness calibration of multichannel sound images; (2) the effectiveness of distance control; (3) the resolution of distance control provided by the system; (4) the effectiveness of the proposed system when compared to a commercially available multichannel room simulation system in terms of control of source distance and proximity to a room boundary; (5) the role of tracking early reflection patterns on the perception of sound source distance; (6) the role of tracking early reflection patterns on the perception of lateral phantom images. The listening tests confirm the effectiveness of the system for control of perceived sound source distance, proximity to room boundaries, and azimuth, through fine, dynamic adjustment of parameters according to source location. All of the parameters are grouped and controlled together to create a perceptually strong impression of source location and movement within a simulated space.
Hu, Hongmei; Kollmeier, Birger; Dietz, Mathias
2016-01-01
Although bilateral cochlear implants (BiCIs) have succeeded in improving the spatial hearing performance of bilateral CI users, the overall performance is still not comparable with normal hearing listeners. Limited success can be partially caused by an interaural mismatch of the place-of-stimulation in each cochlea. Pairing matched interaural CI electrodes and stimulating them with the same frequency band is expected to facilitate binaural functions such as binaural fusion, localization, or spatial release from masking. It has been shown in animal experiments that the magnitude of the binaural interaction component (BIC) derived from the wave-eV decreases for increasing interaural place of stimulation mismatch. This motivated the investigation of the suitability of an electroencephalography-based objective electrode-frequency fitting procedure based on the BIC for BiCI users. A 61 channel monaural and binaural electrically evoked auditory brainstem response (eABR) recording was performed in 7 MED-EL BiCI subjects so far. These BiCI subjects were directly stimulated at 60% dynamic range with 19.9 pulses per second via a research platform provided by the University of Innsbruck (RIB II). The BIC was derived for several interaural electrode pairs by subtracting the response from binaural stimulation from their summed monaural responses. The BIC based pairing results are compared with two psychoacoustic pairing methods: interaural pulse time difference sensitivity and interaural pitch matching. The results for all three methods analyzed as a function of probe electrode allow for determining a matched pair in more than half of the subjects, with a typical accuracy of ± 1 electrode. This includes evidence for statistically significant tuning of the BIC as a function of probe electrode in human subjects. However, results across the three conditions were sometimes not consistent. These discrepancies will be discussed in the light of pitch plasticity versus less plastic brainstem processing.
Neuromagnetic Index of Hemispheric Asymmetry Prognosticating the Outcome of Sudden Hearing Loss
Li, Lieber Po-Hung; Shiao, An-Suey; Chen, Kuang-Chao; Lee, Po-Lei; Niddam, David M.; Chang, Shyue-Yih; Hsieh, Jen-Chuen
2012-01-01
The longitudinal relationship between central plastic changes and clinical presentations of peripheral hearing impairment remains unknown. Previously, we reported a unique plastic pattern of “healthy-side dominance” in acute unilateral idiopathic sudden sensorineural hearing loss (ISSNHL). This study aimed to explore whether such hemispheric asymmetry bears any prognostic relevance to ISSNHL along the disease course. Using magnetoencephalography (MEG), inter-hemispheric differences in peak dipole amplitude and latency of N100m to monaural tones were evaluated in 21 controls and 21 ISSNHL patients at two stages: initial and fixed stage (1 month later). Dynamics/Prognostication of hemispheric asymmetry were assessed by the interplay between hearing level/hearing gain and ipsilateral/contralateral ratio (I/C) of N100m latency and amplitude. Healthy-side dominance of N100m amplitude was observed in ISSNHL initially. The pattern changed with disease process. There is a strong correlation between the hearing level at the fixed stage and initial I/Camplitude on affected-ear stimulation in ISSNHL. The optimal cut-off value with the best prognostication effect for the hearing improvement at the fixed stage was an initial I/Clatency on affected-ear stimulation of 1.34 (between subgroups of complete and partial recovery) and an initial I/Clatency on healthy-ear stimulation of 0.76 (between subgroups of partial and no recovery), respectively. This study suggested that a dynamic process of central auditory plasticity can be induced by peripheral lesions. The hemispheric asymmetry at the initial stage bears an excellent prognostic potential for the treatment outcomes and hearing level at the fixed stage in ISSNHL. Our study demonstrated that such brain signature of central auditory plasticity in terms of both N100m latency and amplitude at defined time can serve as a prognostication predictor for ISSNHL. Further studies are needed to explore the long-term temporal scenario of auditory hemispheric asymmetry and to get better psychoacoustic correlates of pathological hemispheric asymmetry in ISSNHL. PMID:22532839
Computer-automated tinnitus assessment: noise-band matching, maskability, and residual inhibition.
Henry, James A; Roberts, Larry E; Ellingson, Roger M; Thielman, Emily J
2013-06-01
Psychoacoustic measures of tinnitus typically include loudness and pitch match, minimum masking level (MML), and residual inhibition (RI). We previously developed and documented a computer-automated tinnitus evaluation system (TES) capable of subject-guided loudness and pitch matching. The TES was further developed to conduct computer-aided, subject-guided testing for noise-band matching (NBM), MML, and RI. The purpose of the present study was to document the capability of the upgraded TES to obtain measures of NBM, MML, and RI, and to determine the test-retest reliability of the responses obtained. Three subject-guided, computer-automated testing protocols were developed to conduct NBM. For MML and RI testing, a 2-12 kHz band of noise was used. All testing was repeated during a second session. Subjects meeting study criteria were selected from those who had previously been tested for loudness and pitch matching in our laboratory. A total of 21 subjects completed testing, including seven females and 14 males. The upgraded TES was found to be fairly time efficient. Subjects were generally reliable, both within and between sessions, with respect to the type of stimulus they chose as the best match to their tinnitus. Matching to bandwidth was more variable between measurements, with greater consistency seen for subjects reporting tonal tinnitus or wide-band noisy tinnitus than intermediate types. Between-session repeated MMLs were within 10 dB of each other for all but three of the subjects. Subjects who experienced RI during Session 1 tended to be those who experienced it during Session 2. This study may represent the first time that NBM, MML, and RI audiometric testing results have been obtained entirely through a self-contained, computer-automated system designed specifically for use in the clinic. Future plans include refinements to achieve greater testing efficiency. American Academy of Audiology.
Relationship between behavioral and physiological spectral-ripple discrimination.
Won, Jong Ho; Clinard, Christopher G; Kwon, Seeyoun; Dasika, Vasant K; Nie, Kaibao; Drennan, Ward R; Tremblay, Kelly L; Rubinstein, Jay T
2011-06-01
Previous studies have found a significant correlation between spectral-ripple discrimination and speech and music perception in cochlear implant (CI) users. This relationship could be of use to clinicians and scientists who are interested in using spectral-ripple stimuli in the assessment and habilitation of CI users. However, previous psychoacoustic tasks used to assess spectral discrimination are not suitable for all populations, and it would be beneficial to develop methods that could be used to test all age ranges, including pediatric implant users. Additionally, it is important to understand how ripple stimuli are processed in the central auditory system and how their neural representation contributes to behavioral performance. For this reason, we developed a single-interval, yes/no paradigm that could potentially be used both behaviorally and electrophysiologically to estimate spectral-ripple threshold. In experiment 1, behavioral thresholds obtained using the single-interval method were compared to thresholds obtained using a previously established three-alternative forced-choice method. A significant correlation was found (r = 0.84, p = 0.0002) in 14 adult CI users. The spectral-ripple threshold obtained using the new method also correlated with speech perception in quiet and noise. In experiment 2, the effect of the number of vocoder-processing channels on the behavioral and physiological threshold in normal-hearing listeners was determined. Behavioral thresholds, using the new single-interval method, as well as cortical P1-N1-P2 responses changed as a function of the number of channels. Better behavioral and physiological performance (i.e., better discrimination ability at higher ripple densities) was observed as more channels added. In experiment 3, the relationship between behavioral and physiological data was examined. Amplitudes of the P1-N1-P2 "change" responses were significantly correlated with d' values from the single-interval behavioral procedure. Results suggest that the single-interval procedure with spectral-ripple phase inversion in ongoing stimuli is a valid approach for measuring behavioral or physiological spectral resolution.
Zinc supplementation for tinnitus.
Person, Osmar C; Puga, Maria Es; da Silva, Edina Mk; Torloni, Maria R
2016-11-23
Tinnitus is the perception of sound without external acoustic stimuli. Patients with severe tinnitus may have physical and psychological complaints and their tinnitus can cause deterioration in their quality of life. At present no specific therapy for tinnitus has been found to be satisfactory in all patients. In recent decades, a number of reports have suggested that oral zinc supplementation may be effective in the management of tinnitus. Since zinc has a role in cochlear physiology and in the synapses of the auditory system, there is a plausible mechanism of action for this treatment. To evaluate the effectiveness and safety of oral zinc supplementation in the management of patients with tinnitus. The Cochrane ENT Information Specialist searched the ENT Trials Register; Central Register of Controlled Trials (CENTRAL 2016, Issue 6); PubMed; EMBASE; CINAHL; Web of Science; ClinicalTrials.gov; ICTRP and additional sources for published and unpublished trials. The date of the search was 14 July 2016. Randomised controlled trials comparing zinc supplementation versus placebo in adults (18 years and over) with tinnitus. We used the standard methodological procedures recommended by Cochrane. Our primary outcome measures were improvement in tinnitus severity and disability, measured by a validated tinnitus-specific questionnaire, and adverse effects. Secondary outcomes were quality of life, change in socioeconomic impact associated with work, change in anxiety and depression disorders, change in psychoacoustic parameters, change in tinnitus loudness, change in overall severity of tinnitus and change in thresholds on pure tone audiometry. We used GRADE to assess the quality of the evidence for each outcome; this is indicated in italics. We included three trials involving a total of 209 participants. The studies were at moderate to high risk of bias. All included studies had differences in participant selection criteria, length of follow-up and outcome measurement, precluding a meta-analysis. The participants were all adults over 18 years with subjective tinnitus, but one study conducted in 2013 (n = 109) included only elderly patients. Improvement in tinnitus severity and disabilityOnly the study in elderly patients used a validated instrument (Tinnitus Handicap Questionnaire) for this primary outcome. The authors of this cross-over study did not report the results of the two phases separately and found no significant differences in the proportion of patients reporting tinnitus improvement at four months of follow-up: 5% (5/93) versus 2% (2/94) in the zinc and placebo groups, respectively (risk ratio (RR) 2.53, 95% confidence interval (CI) 0.50 to 12.70; very low-quality evidence).None of the included studies reported any significant adverse effects. Secondary outcomesFor the secondary outcome change in tinnitus loudness, one study reported no significant difference between the zinc and placebo groups after eight weeks: mean difference in tinnitus loudness -9.71 dB (95% CI -25.53 to 6.11; very low-quality evidence). Another study also measured tinnitus loudness but used a 0- to 100-point scale. The authors of this second study reported no significant difference between the zinc and placebo groups after four months: mean difference in tinnitus loudness rating scores 0.50 (95% CI -5.08 to 6.08; very low-quality evidence).Two studies used unvalidated instruments to assess tinnitus severity. One (with 50 participants) reported the severity of tinnitus using a non-validated scale (0 to 7 points) and found no significant difference in subjective tinnitus scores between the zinc and placebo groups at the end of eight weeks of follow-up (mean difference (MD) -1.41, 95% CI -2.97 to 0.15; very low-quality evidence). A third trial (n = 50) also evaluated the improvement of tinnitus using a non-validated instrument (a 0 to 10 scale: 10 = severe and unbearable tinnitus). In this study, after eight weeks there was no difference in the proportion of patients with improvement in their tinnitus, 8.7% (2/23) treated with zinc versus 8% (2/25) of those who received a placebo (RR 1.09, 95% CI 0.17 to 7.10, very low-quality evidence).None of the included studies reported any of our other secondary outcomes (quality of life, change in socioeconomic impact associated with work, change in anxiety and depression disorders, change in psychoacoustic parameters or change in thresholds on pure tone audiometry). We found no evidence that the use of oral zinc supplementation improves symptoms in adults with tinnitus.
Durai, Mithila; O'Keeffe, Mary G; Searchfield, Grant D
2017-03-01
Existing evidence suggests a strong relationship between tinnitus and emotion. The objective of this study was to examine the effects of short-term emotional changes along valence and arousal dimensions on tinnitus outcomes. Emotional stimuli were presented in two different modalities: auditory and visual. The authors hypothesized that (1) negative valence (unpleasant) stimuli and/or high arousal stimuli will lead to greater tinnitus loudness and annoyance than positive valence and/or low arousal stimuli, and (2) auditory emotional stimuli, which are in the same modality as the tinnitus, will exhibit a greater effect on tinnitus outcome measures than visual stimuli. Auditory and visual emotive stimuli were administered to 22 participants (12 females and 10 males) with chronic tinnitus, recruited via email invitations send out to the University of Auckland Tinnitus Research Volunteer Database. Emotional stimuli used were taken from the International Affective Digital Sounds- Version 2 (IADS-2) and the International Affective Picture System (IAPS) (Bradley and Lang, 2007a, 2007b). The Emotion Regulation Questionnaire (Gross and John, 2003) was administered alongside subjective ratings of tinnitus loudness and annoyance, and psychoacoustic sensation level matches to external sounds. Males had significantly different emotional regulation scores than females. Negative valence emotional auditory stimuli led to higher tinnitus loudness ratings in males and females and higher annoyance ratings in males only; loudness matches of tinnitus remained unchanged. The visual stimuli did not have an effect on tinnitus ratings. The results are discussed relative to the Adaptation Level Theory Model of Tinnitus. The results indicate that the negative valence dimension of emotion is associated with increased tinnitus magnitude judgements and gender effects may also be present, but only when the emotional stimulus is in the auditory modality. Sounds with emotional associations may be used for sound therapy for tinnitus relief; it is of interest to determine whether the emotional component of sound treatments can play a role in reversing the negative responses discussed in this paper. Copyright © 2016 Elsevier B.V. All rights reserved.
Achieving perceptually-accurate aural telepresence
NASA Astrophysics Data System (ADS)
Henderson, Paul D.
Immersive multimedia requires not only realistic visual imagery but also a perceptually-accurate aural experience. A sound field may be presented simultaneously to a listener via a loudspeaker rendering system using the direct sound from acoustic sources as well as a simulation or "auralization" of room acoustics. Beginning with classical Wave-Field Synthesis (WFS), improvements are made to correct for asymmetries in loudspeaker array geometry. Presented is a new Spatially-Equalized WFS (SE-WFS) technique to maintain the energy-time balance of a simulated room by equalizing the reproduced spectrum at the listener for a distribution of possible source angles. Each reproduced source or reflection is filtered according to its incidence angle to the listener. An SE-WFS loudspeaker array of arbitrary geometry reproduces the sound field of a room with correct spectral and temporal balance, compared with classically-processed WFS systems. Localization accuracy of human listeners in SE-WFS sound fields is quantified by psychoacoustical testing. At a loudspeaker spacing of 0.17 m (equivalent to an aliasing cutoff frequency of 1 kHz), SE-WFS exhibits a localization blur of 3 degrees, nearly equal to real point sources. Increasing the loudspeaker spacing to 0.68 m (for a cutoff frequency of 170 Hz) results in a blur of less than 5 degrees. In contrast, stereophonic reproduction is less accurate with a blur of 7 degrees. The ventriloquist effect is psychometrically investigated to determine the effect of an intentional directional incongruence between audio and video stimuli. Subjects were presented with prerecorded full-spectrum speech and motion video of a talker's head as well as broadband noise bursts with a static image. The video image was displaced from the audio stimulus in azimuth by varying amounts, and the perceived auditory location measured. A strong bias was detectable for small angular discrepancies between audio and video stimuli for separations of less than 8 degrees for speech and less than 4 degrees with a pink noise burst. The results allow for the density of WFS systems to be selected from the required localization accuracy. Also, by exploiting the ventriloquist effect, the angular resolution of an audio rendering may be reduced when combined with spatially-accurate video.
Incorporating Auditory Models in Speech/Audio Applications
NASA Astrophysics Data System (ADS)
Krishnamoorthi, Harish
2011-12-01
Following the success in incorporating perceptual models in audio coding algorithms, their application in other speech/audio processing systems is expanding. In general, all perceptual speech/audio processing algorithms involve minimization of an objective function that directly/indirectly incorporates properties of human perception. This dissertation primarily investigates the problems associated with directly embedding an auditory model in the objective function formulation and proposes possible solutions to overcome high complexity issues for use in real-time speech/audio algorithms. Specific problems addressed in this dissertation include: 1) the development of approximate but computationally efficient auditory model implementations that are consistent with the principles of psychoacoustics, 2) the development of a mapping scheme that allows synthesizing a time/frequency domain representation from its equivalent auditory model output. The first problem is aimed at addressing the high computational complexity involved in solving perceptual objective functions that require repeated application of auditory model for evaluation of different candidate solutions. In this dissertation, a frequency pruning and a detector pruning algorithm is developed that efficiently implements the various auditory model stages. The performance of the pruned model is compared to that of the original auditory model for different types of test signals in the SQAM database. Experimental results indicate only a 4-7% relative error in loudness while attaining up to 80-90 % reduction in computational complexity. Similarly, a hybrid algorithm is developed specifically for use with sinusoidal signals and employs the proposed auditory pattern combining technique together with a look-up table to store representative auditory patterns. The second problem obtains an estimate of the auditory representation that minimizes a perceptual objective function and transforms the auditory pattern back to its equivalent time/frequency representation. This avoids the repeated application of auditory model stages to test different candidate time/frequency vectors in minimizing perceptual objective functions. In this dissertation, a constrained mapping scheme is developed by linearizing certain auditory model stages that ensures obtaining a time/frequency mapping corresponding to the estimated auditory representation. This paradigm was successfully incorporated in a perceptual speech enhancement algorithm and a sinusoidal component selection task.
Investigations in mechanisms and strategies to enhance hearing with cochlear implants
NASA Astrophysics Data System (ADS)
Churchill, Tyler H.
Cochlear implants (CIs) produce hearing sensations by stimulating the auditory nerve (AN) with current pulses whose amplitudes are modulated by filtered acoustic temporal envelopes. While this technology has provided hearing for multitudinous CI recipients, even bilaterally-implanted listeners have more difficulty understanding speech in noise and localizing sounds than normal hearing (NH) listeners. Three studies reported here have explored ways to improve electric hearing abilities. Vocoders are often used to simulate CIs for NH listeners. Study 1 was a psychoacoustic vocoder study examining the effects of harmonic carrier phase dispersion and simulated CI current spread on speech intelligibility in noise. Results showed that simulated current spread was detrimental to speech understanding and that speech vocoded with carriers whose components' starting phases were equal was the least intelligible. Cross-correlogram analyses of AN model simulations confirmed that carrier component phase dispersion resulted in better neural envelope representation. Localization abilities rely on binaural processing mechanisms in the brainstem and mid-brain that are not fully understood. In Study 2, several potential mechanisms were evaluated based on the ability of metrics extracted from stereo AN simulations to predict azimuthal locations. Results suggest that unique across-frequency patterns of binaural cross-correlation may provide a strong cue set for lateralization and that interaural level differences alone cannot explain NH sensitivity to lateral position. While it is known that many bilateral CI users are sensitive to interaural time differences (ITDs) in low-rate pulsatile stimulation, most contemporary CI processing strategies use high-rate, constant-rate pulse trains. In Study 3, we examined the effects of pulse rate and pulse timing on ITD discrimination, ITD lateralization, and speech recognition by bilateral CI listeners. Results showed that listeners were able to use low-rate pulse timing cues presented redundantly on multiple electrodes for ITD discrimination and lateralization of speech stimuli even when mixed with high rates on other electrodes. These results have contributed to a better understanding of those aspects of the auditory system that support speech understanding and binaural hearing, suggested vocoder parameters that may simulate aspects of electric hearing, and shown that redundant, low-rate pulse timing supports improved spatial hearing for bilateral CI listeners.
Loudness of dynamic stimuli in acoustic and electric hearing.
Zhang, C; Zeng, F G
1997-11-01
Traditional loudness models have been based on the average energy and the critical band analysis of steady-state sounds. However, most environmental sounds, including speech, are dynamic stimuli, in which the average level [e.g., the root-mean-square (rms) level] does not account for the large temporal fluctuations. The question addressed here was whether two stimuli of the same rms level but different peak levels would produce an equal loudness sensation. A modern adaptive procedure was used to replicate two classic experiments demonstrating that the sensation of "beats" in a two- or three-tone complex resulted in a louder sensation [E. Zwicker and H. Fastl, Psychoacoustics-Facts and Models (Springer-Verlag, Berlin, 1990)]. Two additional experiments were conducted to study exclusively the effects of the temporal envelope on the loudness sensation of dynamic stimuli. Loudness balance was performed by normal-hearing listeners between a white noise and a sinusoidally amplitude-modulated noise in one experiment, and by cochlear implant listeners between two harmonic stimuli of the same magnitude spectra, but different phase spectra, in the other experiment. The results from both experiments showed that, for two stimuli of the same rms level, the stimulus with greater temporal fluctuations sometimes produced a significantly louder sensation, depending on the temporal frequency and overall stimulus level. In normal-hearing listeners, the louder sensation was produced for the amplitude-modulated stimuli with modulation frequencies lower than 400 Hz, and gradually disappeared above 400 Hz, resulting in a low-pass filtering characteristic which bore some similarity to the temporal modulation transfer function. The extent to which loudness was greater was a nonmonotonic function of level in acoustic hearing and a monotonically increasingly function in electric hearing. These results suggest that the loudness sensation of a dynamic stimulus is not limited to a 100-ms temporal integration process, and may be determined jointly by a compression process in the cochlea and an expansion process in the brain. A level-dependent compression scheme that may better restore normal loudness of dynamic stimuli in hearing aids and cochlear implants is proposed.
A model of head-related transfer functions based on a state-space analysis
NASA Astrophysics Data System (ADS)
Adams, Norman Herkamp
This dissertation develops and validates a novel state-space method for binaural auditory display. Binaural displays seek to immerse a listener in a 3D virtual auditory scene with a pair of headphones. The challenge for any binaural display is to compute the two signals to supply to the headphones. The present work considers a general framework capable of synthesizing a wide variety of auditory scenes. The framework models collections of head-related transfer functions (HRTFs) simultaneously. This framework improves the flexibility of contemporary displays, but it also compounds the steep computational cost of the display. The cost is reduced dramatically by formulating the collection of HRTFs in the state-space and employing order-reduction techniques to design efficient approximants. Order-reduction techniques based on the Hankel-operator are found to yield accurate low-cost approximants. However, the inter-aural time difference (ITD) of the HRTFs degrades the time-domain response of the approximants. Fortunately, this problem can be circumvented by employing a state-space architecture that allows the ITD to be modeled outside of the state-space. Accordingly, three state-space architectures are considered. Overall, a multiple-input, single-output (MISO) architecture yields the best compromise between performance and flexibility. The state-space approximants are evaluated both empirically and psychoacoustically. An array of truncated FIR filters is used as a pragmatic reference system for comparison. For a fixed cost bound, the state-space systems yield lower approximation error than FIR arrays for D>10, where D is the number of directions in the HRTF collection. A series of headphone listening tests are also performed to validate the state-space approach, and to estimate the minimum order N of indiscriminable approximants. For D = 50, the state-space systems yield order thresholds less than half those of the FIR arrays. Depending upon the stimulus uncertainty, a minimum state-space order of 7≤N≤23 appears to be adequate. In conclusion, the proposed state-space method enables a more flexible and immersive binaural display with low computational cost.
Validation of Online Versions of Tinnitus Questionnaires Translated into Swedish.
Müller, Karolina; Edvall, Niklas K; Idrizbegovic, Esma; Huhn, Robert; Cima, Rilana; Persson, Viktor; Leineweber, Constanze; Westerlund, Hugo; Langguth, Berthold; Schlee, Winfried; Canlon, Barbara; Cederroth, Christopher R
2016-01-01
Background: Due to the lack of objective measures for assessing tinnitus, its clinical evaluation largely relies on the use of questionnaires and psychoacoustic tests. A global assessment of tinnitus burden would largely benefit from holistic approaches that not only incorporate measures of tinnitus but also take into account associated fears, emotional aspects (stress, anxiety, and depression), and quality of life. In Sweden, only a few instruments are available for assessing tinnitus, and the existing tools lack validation. Therefore, we translated a set of questionnaires into Swedish and evaluated their reliability and validity in a group of tinnitus subjects. Methods: We translated the English versions of the Tinnitus Functional Index (TFI), the Fear of Tinnitus Questionnaire (FTQ), the Tinnitus Catastrophizing Scale (TCS), the Perceived Stress Questionnaire (PSQ-30), and the Tinnitus Sample Case History Questionnaire (TSCHQ) into Swedish. These translations were delivered via the internet with the already existing Swedish versions of the Tinnitus Handicap Inventory (THI), the Hospital Anxiety and Depression Scale (HADS), the Hyperacusis Questionnaire (HQ), and the World Health Organization Quality of Life questionnaire (WHOQoL-BREF). Psychometric properties were evaluated by means of internal consistency [Cronbach's alpha (α)] and test-retest reliability across a 9-week interval [Intraclass Correlation Coefficient (ICC), Cohen's kappa] in order to establish construct as well as clinical validity using a sample of 260 subjects from a population-based cohort. Results: Internal consistency was acceptable for all questionnaires (α > 0.7) with the exception of the "social relationships" subscale of the WHOQoL-BREF. Test-retest reliability was generally acceptable (ICC > 0.70, Cohens kappa > 0.60) for the tinnitus-related questionnaires, except for the TFI "sense of control" subscale and 15 items of the TSCHQ. Spearmen rank correlations showed that almost all questionnaires on tinnitus are significantly related, indicating that these questionnaires measure different aspects of the same construct. The data supported good clinical validity of the tinnitus-related questionnaires. Conclusion: Our results suggest that most Swedish adaptations of the questionnaires are suitable for clinical and research settings and should facilitate the assessment of treatment outcomes using a more holistic approach by including measures of tinnitus fears, emotional burden, and quality of life.
The Impact of Single-Sided Deafness upon Music Appreciation.
Meehan, Sarah; Hough, Elizabeth A; Crundwell, Gemma; Knappett, Rachel; Smith, Mark; Baguley, David M
2017-05-01
Many of the world's population have hearing loss in one ear; current statistics indicate that up to 10% of the population may be affected. Although the detrimental impact of bilateral hearing loss, hearing aids, and cochlear implants upon music appreciation is well recognized, studies on the influence of single-sided deafness (SSD) are sparse. We sought to investigate whether a single-sided hearing loss can cause problems with music appreciation, despite normal hearing in the other ear. A tailored questionnaire was used to investigate music appreciation for those with SSD. We performed a retrospective survey of a population of 51 adults from a University Hospital Audiology Department SSD clinic. SSD was predominantly adult-onset sensorineural hearing loss, caused by a variety of etiologies. Analyses were performed to assess for statistical differences between groups, for example, comparing music appreciation before and after the onset of SSD, or before and after receiving hearing aid(s). Results demonstrated that a proportion of the population experienced significant changes to the way music sounded; music was found to sound more unnatural (75%), unpleasant (71%), and indistinct (81%) than before hearing loss. Music was reported to lack the perceptual qualities of stereo sound, and to be confounded by distortion effects and tinnitus. Such changes manifested in an altered music appreciation, with 44% of participants listening to music less often, 71% of participants enjoying music less, and 46% of participants reporting that music played a lesser role in their lives than pre-SSD. Negative effects surrounding social occasions with music were revealed, along with a strong preference for limiting background music. Hearing aids were not found to significantly ameliorate these effects. Results could be explained in part through considerations of psychoacoustic changes intrinsic to an asymmetric hearing loss and impaired auditory scene analysis. Given the prevalence of music and its capacity to influence an individual's well-being, results here present strong indications that the potential effects of SSD on music appreciation should be considered in a clinical context; an investigation into relevant rehabilitation techniques may prove valuable. American Academy of Audiology
Validation of Online Versions of Tinnitus Questionnaires Translated into Swedish
Müller, Karolina; Edvall, Niklas K.; Idrizbegovic, Esma; Huhn, Robert; Cima, Rilana; Persson, Viktor; Leineweber, Constanze; Westerlund, Hugo; Langguth, Berthold; Schlee, Winfried; Canlon, Barbara; Cederroth, Christopher R.
2016-01-01
Background: Due to the lack of objective measures for assessing tinnitus, its clinical evaluation largely relies on the use of questionnaires and psychoacoustic tests. A global assessment of tinnitus burden would largely benefit from holistic approaches that not only incorporate measures of tinnitus but also take into account associated fears, emotional aspects (stress, anxiety, and depression), and quality of life. In Sweden, only a few instruments are available for assessing tinnitus, and the existing tools lack validation. Therefore, we translated a set of questionnaires into Swedish and evaluated their reliability and validity in a group of tinnitus subjects. Methods: We translated the English versions of the Tinnitus Functional Index (TFI), the Fear of Tinnitus Questionnaire (FTQ), the Tinnitus Catastrophizing Scale (TCS), the Perceived Stress Questionnaire (PSQ-30), and the Tinnitus Sample Case History Questionnaire (TSCHQ) into Swedish. These translations were delivered via the internet with the already existing Swedish versions of the Tinnitus Handicap Inventory (THI), the Hospital Anxiety and Depression Scale (HADS), the Hyperacusis Questionnaire (HQ), and the World Health Organization Quality of Life questionnaire (WHOQoL-BREF). Psychometric properties were evaluated by means of internal consistency [Cronbach's alpha (α)] and test–retest reliability across a 9-week interval [Intraclass Correlation Coefficient (ICC), Cohen's kappa] in order to establish construct as well as clinical validity using a sample of 260 subjects from a population-based cohort. Results: Internal consistency was acceptable for all questionnaires (α > 0.7) with the exception of the “social relationships” subscale of the WHOQoL-BREF. Test–retest reliability was generally acceptable (ICC > 0.70, Cohens kappa > 0.60) for the tinnitus-related questionnaires, except for the TFI “sense of control” subscale and 15 items of the TSCHQ. Spearmen rank correlations showed that almost all questionnaires on tinnitus are significantly related, indicating that these questionnaires measure different aspects of the same construct. The data supported good clinical validity of the tinnitus-related questionnaires. Conclusion: Our results suggest that most Swedish adaptations of the questionnaires are suitable for clinical and research settings and should facilitate the assessment of treatment outcomes using a more holistic approach by including measures of tinnitus fears, emotional burden, and quality of life. PMID:27920720
Vernooij, Eveline; Orcalli, Angelo; Fabbro, Franco; Crescentini, Cristiano
2016-01-01
The endless scale illusion, obtained by cyclically repeating a chromatic scale made up of Shepard tones, has been used in a variety of musical works. Music psychology and neuroscience has been interested in this particular psychoacoustic phenomenon mainly for studying the cognitive processes of pitch perception involved. In the present study, we investigated the emotional states induced by the Shepard-Risset glissando, a variant of the Shepard scale. For this purpose we chose three musical stimuli: a Matlab-generated Shepard Risset glissando, Jean-Claude Risset's Computer Suite from Little Boy, which presents a Shepard-Risset glissando integrated in the aesthetic context of a composition, and an ordinary orchestral glissando taken from the opening of Iannis Xenakis's Metastasis. Seventy-three volunteers completed a listening experiment during which they rated their emotional response to these stimuli on a seven-point Likert scale and indicated whether they had experienced a disruption of equilibrium. Personality was also measured with the Five-Factor Model of personality traits. The results show that negative emotions were most strongly evoked during listening to each of the stimuli. We also found that the Shepard-Risset glissando illusion, both within the aesthetic context of a musical composition and on its own, was capable of evoking disruption of equilibrium, frequently leading to the associated feeling of falling. Moreover, generally for the Shepard-Risset glissando illusion, higher negative emotional ratings were given by individuals who had experienced a feeling of disturbance of equilibrium relative to those who had not had this experience. Finally, we found a complex pattern of relationships between personality and the subjective experience of the glissando. Openness to experience correlated positively with positive emotion ratings for the Computer Suite, while agreeableness correlated negatively with positive emotion ratings for the Matlab stimulus. Moreover, results indicated higher (Bonferroni-uncorrected) neuroticism for those who experienced an equilibrium disturbance relative to subjects who did not have this experience during listening to the Computer Suite. These findings suggest that musical paradoxes may be of interest not only for the insights they provide on our perceptual system, but also for the richness of the emotional experience elicited during listening. PMID:26973584
NASA Astrophysics Data System (ADS)
Walls, Kimberly Kyle Curley
1992-01-01
Musical expression is largely dependent upon accentuation, yet there have been few attempts to study the perception of dynamic accent in music or to relate the results of psychoacoustical research in intensity to realistic musical situations. The purpose of the experiment was to estimate the relationships among (a) the intensity increment in dB(A) required to meet an 80% correct criterion in the perception of one accented tone embedded within a seven -tone isochronous series of identical 87 dB(A) snare drum timbre stimuli of 333 ms onsets (accent level, or AL), (b) the different limen (DL) for intensity increase to meet a 75% correct criterion in a 2AFC task for pairs for the stimuli, and (c) the age of the subjects, all of whom have normal audiograms. The 51 subjects (N = 51) were female nonmusicians ranging in age from 9 to 33 years (M = 17.98, SD = 5.21). The response tasks involved saying whether the second tone of each pair was louder or softer and circling the accented note in notated quarter notes. The stimuli production, the headphone calibration process, and their rationales were detailed. The global regression model was significant (F(2, 48) = 5.505, p =.007, R^2 =.187), and the relationship between AL and DL was not significant (F(1, 48) = 5.505, p =.197, R^2 change =.029), the relationship between AL and age was significant (F(1, 48) = 5.732, p =.021, R ^2 change =.098) at an alpha level of.05 and power calculated at.66 for a medium ES. It was concluded that accented sounds are easier to perceive in tone pairs than they are in a musical setting and that subject maturation improves performance of intensity judgement tasks. Suggestions for further research include shortening the length of the experimental session for younger subjects and increasing the number of intensity increments as well as using smaller increments to accommodate individual differences in perception.
Human-based percussion and self-similarity detection in electroacoustic music
NASA Astrophysics Data System (ADS)
Mills, John Anderson, III
Electroacoustic music is music that uses electronic technology for the compositional manipulation of sound, and is a unique genre of music for many reasons. Analyzing electroacoustic music requires special measures, some of which are integrated into the design of a preliminary percussion analysis tool set for electroacoustic music. This tool set is designed to incorporate the human processing of music and sound. Models of the human auditory periphery are used as a front end to the analysis algorithms. The audio properties of percussivity and self-similarity are chosen as the focus because these properties are computable and informative. A collection of human judgments about percussion was undertaken to acquire clearly specified, sound-event dimensions that humans use as a percussive cue. A total of 29 participants was asked to make judgments about the percussivity of 360 pairs of synthesized snare-drum sounds. The grouped results indicate that of the dimensions tested rise time is the strongest cue for percussivity. String resonance also has a strong effect, but because of the complex nature of string resonance, it is not a fundamental dimension of a sound event. Gross spectral filtering also has an effect on the judgment of percussivity but the effect is weaker than for rise time and string resonance. Gross spectral filtering also has less effect when the stronger cue of rise time is modified simultaneously. A percussivity-profile algorithm (PPA) is designed to identify those instants in pieces of music that humans also would identify as percussive. The PPA is implemented using a time-domain, channel-based approach and psychoacoustic models. The input parameters are tuned to maximize performance at matching participants' choices in the percussion-judgment collection. After the PPA is tuned, the PPA then is used to analyze pieces of electroacoustic music. Real electroacoustic music introduces new challenges for the PPA, though those same challenges might affect human judgment as well. A similarity matrix is combined with the PPA in order to find self-similarity in the percussive sounds of electroacoustic music. This percussive similarity matrix is then used to identify structural characteristics in two pieces of electroacoustic music.
Neurometric amplitude-modulation detection threshold in the guinea-pig ventral cochlear nucleus
Sayles, Mark; Füllgrabe, Christian; Winter, Ian M
2013-01-01
Amplitude modulation (AM) is a pervasive feature of natural sounds. Neural detection and processing of modulation cues is behaviourally important across species. Although most ecologically relevant sounds are not fully modulated, physiological studies have usually concentrated on fully modulated (100% modulation depth) signals. Psychoacoustic experiments mainly operate at low modulation depths, around detection threshold (∼5% AM). We presented sinusoidal amplitude-modulated tones, systematically varying modulation depth between zero and 100%, at a range of modulation frequencies, to anaesthetised guinea-pigs while recording spikes from neurons in the ventral cochlear nucleus (VCN). The cochlear nucleus is the site of the first synapse in the central auditory system. At this locus significant signal processing occurs with respect to representation of AM signals. Spike trains were analysed in terms of the vector strength of spike synchrony to the amplitude envelope. Neurons showed either low-pass or band-pass temporal modulation transfer functions, with the proportion of band-pass responses increasing with increasing sound level. The proportion of units showing a band-pass response varies with unit type: sustained chopper (CS) > transient chopper (CT) > primary-like (PL). Spike synchrony increased with increasing modulation depth. At the lowest modulation depth (6%), significant spike synchrony was only observed near to the unit's best modulation frequency for all unit types tested. Modulation tuning therefore became sharper with decreasing modulation depth. AM detection threshold was calculated for each individual unit as a function of modulation frequency. Chopper units have significantly better AM detection thresholds than do primary-like units. AM detection threshold is significantly worse at 40 dB vs. 10 dB above pure-tone spike rate threshold. Mean modulation detection thresholds for sounds 10 dB above pure-tone spike rate threshold at best modulation frequency are (95% CI) 11.6% (10.0–13.1) for PL units, 9.8% (8.2–11.5) for CT units, and 10.8% (8.4–13.2) for CS units. The most sensitive guinea-pig VCN single unit AM detection thresholds are similar to human psychophysical performance (∼3% AM), while the mean neurometric thresholds approach whole animal behavioural performance (∼10% AM). PMID:23629508
Vernooij, Eveline; Orcalli, Angelo; Fabbro, Franco; Crescentini, Cristiano
2016-01-01
The endless scale illusion, obtained by cyclically repeating a chromatic scale made up of Shepard tones, has been used in a variety of musical works. Music psychology and neuroscience has been interested in this particular psychoacoustic phenomenon mainly for studying the cognitive processes of pitch perception involved. In the present study, we investigated the emotional states induced by the Shepard-Risset glissando, a variant of the Shepard scale. For this purpose we chose three musical stimuli: a Matlab-generated Shepard Risset glissando, Jean-Claude Risset's Computer Suite from Little Boy, which presents a Shepard-Risset glissando integrated in the aesthetic context of a composition, and an ordinary orchestral glissando taken from the opening of Iannis Xenakis's Metastasis. Seventy-three volunteers completed a listening experiment during which they rated their emotional response to these stimuli on a seven-point Likert scale and indicated whether they had experienced a disruption of equilibrium. Personality was also measured with the Five-Factor Model of personality traits. The results show that negative emotions were most strongly evoked during listening to each of the stimuli. We also found that the Shepard-Risset glissando illusion, both within the aesthetic context of a musical composition and on its own, was capable of evoking disruption of equilibrium, frequently leading to the associated feeling of falling. Moreover, generally for the Shepard-Risset glissando illusion, higher negative emotional ratings were given by individuals who had experienced a feeling of disturbance of equilibrium relative to those who had not had this experience. Finally, we found a complex pattern of relationships between personality and the subjective experience of the glissando. Openness to experience correlated positively with positive emotion ratings for the Computer Suite, while agreeableness correlated negatively with positive emotion ratings for the Matlab stimulus. Moreover, results indicated higher (Bonferroni-uncorrected) neuroticism for those who experienced an equilibrium disturbance relative to subjects who did not have this experience during listening to the Computer Suite. These findings suggest that musical paradoxes may be of interest not only for the insights they provide on our perceptual system, but also for the richness of the emotional experience elicited during listening.
Efficient audio signal processing for embedded systems
NASA Astrophysics Data System (ADS)
Chiu, Leung Kin
As mobile platforms continue to pack on more computational power, electronics manufacturers start to differentiate their products by enhancing the audio features. However, consumers also demand smaller devices that could operate for longer time, hence imposing design constraints. In this research, we investigate two design strategies that would allow us to efficiently process audio signals on embedded systems such as mobile phones and portable electronics. In the first strategy, we exploit properties of the human auditory system to process audio signals. We designed a sound enhancement algorithm to make piezoelectric loudspeakers sound ”richer" and "fuller." Piezoelectric speakers have a small form factor but exhibit poor response in the low-frequency region. In the algorithm, we combine psychoacoustic bass extension and dynamic range compression to improve the perceived bass coming out from the tiny speakers. We also developed an audio energy reduction algorithm for loudspeaker power management. The perceptually transparent algorithm extends the battery life of mobile devices and prevents thermal damage in speakers. This method is similar to audio compression algorithms, which encode audio signals in such a ways that the compression artifacts are not easily perceivable. Instead of reducing the storage space, however, we suppress the audio contents that are below the hearing threshold, therefore reducing the signal energy. In the second strategy, we use low-power analog circuits to process the signal before digitizing it. We designed an analog front-end for sound detection and implemented it on a field programmable analog array (FPAA). The system is an example of an analog-to-information converter. The sound classifier front-end can be used in a wide range of applications because programmable floating-gate transistors are employed to store classifier weights. Moreover, we incorporated a feature selection algorithm to simplify the analog front-end. A machine learning algorithm AdaBoost is used to select the most relevant features for a particular sound detection application. In this classifier architecture, we combine simple "base" analog classifiers to form a strong one. We also designed the circuits to implement the AdaBoost-based analog classifier.
Vocal development and auditory perception in CBA/CaJ mice
NASA Astrophysics Data System (ADS)
Radziwon, Kelly E.
Mice are useful laboratory subjects because of their small size, their modest cost, and the fact that researchers have created many different strains to study a variety of disorders. In particular, researchers have found nearly 100 naturally occurring mouse mutations with hearing impairments. For these reasons, mice have become an important model for studies of human deafness. Although much is known about the genetic makeup and physiology of the laboratory mouse, far less is known about mouse auditory behavior. To fully understand the effects of genetic mutations on hearing, it is necessary to determine the hearing abilities of these mice. Two experiments here examined various aspects of mouse auditory perception using CBA/CaJ mice, a commonly used mouse strain. The frequency difference limens experiment tested the mouse's ability to discriminate one tone from another based solely on the frequency of the tone. The mice had similar thresholds as wild mice and gerbils but needed a larger change in frequency than humans and cats. The second psychoacoustic experiment sought to determine which cue, frequency or duration, was more salient when the mice had to identify various tones. In this identification task, the mice overwhelmingly classified the tones based on frequency instead of duration, suggesting that mice are using frequency when differentiating one mouse vocalization from another. The other two experiments were more naturalistic and involved both auditory perception and mouse vocal production. Interest in mouse vocalizations is growing because of the potential for mice to become a model of human speech disorders. These experiments traced mouse vocal development from infant to adult, and they tested the mouse's preference for various vocalizations. This was the first known study to analyze the vocalizations of individual mice across development. Results showed large variation in calling rates among the three cages of adult mice but results were highly consistent across all infant vocalizations. Although the preference experiment did not reveal significant differences between various mouse vocalizations, suggestions are given for future attempts to identify mouse preferences for auditory stimuli.
Liang, Chun; Earl, Brian; Thompson, Ivy; Whitaker, Kayla; Cahn, Steven; Xiang, Jing; Fu, Qian-Jie; Zhang, Fawen
2016-01-01
Objective: The objectives of this study were: (1) to determine if musicians have a better ability to detect frequency changes under quiet and noisy conditions; (2) to use the acoustic change complex (ACC), a type of electroencephalographic (EEG) response, to understand the neural substrates of musician vs. non-musician difference in frequency change detection abilities. Methods: Twenty-four young normal hearing listeners (12 musicians and 12 non-musicians) participated. All participants underwent psychoacoustic frequency detection tests with three types of stimuli: tones (base frequency at 160 Hz) containing frequency changes (Stim 1), tones containing frequency changes masked by low-level noise (Stim 2), and tones containing frequency changes masked by high-level noise (Stim 3). The EEG data were recorded using tones (base frequency at 160 and 1200 Hz, respectively) containing different magnitudes of frequency changes (0, 5, and 50% changes, respectively). The late-latency evoked potential evoked by the onset of the tones (onset LAEP or N1-P2 complex) and that evoked by the frequency change contained in the tone (the acoustic change complex or ACC or N1′-P2′ complex) were analyzed. Results: Musicians significantly outperformed non-musicians in all stimulus conditions. The ACC and onset LAEP showed similarities and differences. Increasing the magnitude of frequency change resulted in increased ACC amplitudes. ACC measures were found to be significantly different between musicians (larger P2′ amplitude) and non-musicians for the base frequency of 160 Hz but not 1200 Hz. Although the peak amplitude in the onset LAEP appeared to be larger and latency shorter in musicians than in non-musicians, the difference did not reach statistical significance. The amplitude of the onset LAEP is significantly correlated with that of the ACC for the base frequency of 160 Hz. Conclusion: The present study demonstrated that musicians do perform better than non-musicians in detecting frequency changes in quiet and noisy conditions. The ACC and onset LAEP may involve different but overlapping neural mechanisms. Significance: This is the first study using the ACC to examine music-training effects. The ACC measures provide an objective tool for documenting musical training effects on frequency detection. PMID:27826221
Making sense of listening: the IMAP test battery.
Barry, Johanna G; Ferguson, Melanie A; Moore, David R
2010-10-11
The ability to hear is only the first step towards making sense of the range of information contained in an auditory signal. Of equal importance are the abilities to extract and use the information encoded in the auditory signal. We refer to these as listening skills (or auditory processing AP). Deficits in these skills are associated with delayed language and literacy development, though the nature of the relevant deficits and their causal connection with these delays is hotly debated. When a child is referred to a health professional with normal hearing and unexplained difficulties in listening, or associated delays in language or literacy development, they should ideally be assessed with a combination of psychoacoustic (AP) tests, suitable for children and for use in a clinic, together with cognitive tests to measure attention, working memory, IQ, and language skills. Such a detailed examination needs to be relatively short and within the technical capability of any suitably qualified professional. Current tests for the presence of AP deficits tend to be poorly constructed and inadequately validated within the normal population. They have little or no reference to the presenting symptoms of the child, and typically include a linguistic component. Poor performance may thus reflect problems with language rather than with AP. To assist in the assessment of children with listening difficulties, pediatric audiologists need a single, standardized child-appropriate test battery based on the use of language-free stimuli. We present the IMAP test battery which was developed at the MRC Institute of Hearing Research to supplement tests currently used to investigate cases of suspected AP deficits. IMAP assesses a range of relevant auditory and cognitive skills and takes about one hour to complete. It has been standardized in 1500 normally-hearing children from across the UK, aged 6-11 years. Since its development, it has been successfully used in a number of large scale studies both in the UK and the USA. IMAP provides measures for separating out sensory from cognitive contributions to hearing. It further limits confounds due to procedural effects by presenting tests in a child-friendly game-format. Stimulus-generation, management of test protocols and control of test presentation is mediated by the IHR-STAR software platform. This provides a standardized methodology for a range of applications and ensures replicable procedures across testers. IHR-STAR provides a flexible, user-programmable environment that currently has additional applications for hearing screening, mapping cochlear implant electrodes, and academic research or teaching.
Sonification of acoustic emission data
NASA Astrophysics Data System (ADS)
Raith, Manuel; Große, Christian
2014-05-01
While loading different specimens, acoustic emissions appear due to micro crack formation or friction of already existing crack edges. These acoustic emissions can be recorded using suitable ultrasonic transducers and transient recorders. The analysis of acoustic emissions can be used to investigate the mechanical behavior of different specimens under load. Our working group has undertaken several experiments, monitored with acoustic emission techniques. Different materials such as natural stone, concrete, wood, steel, carbon composites and bone were investigated. Also the experimental setup has been varied. Fire-spalling experiments on ultrahigh performance concrete and pullout experiments on bonded anchors have been carried out. Furthermore uniaxial compression tests on natural stone and animal bone had been conducted. The analysis tools include not only the counting of events but the analysis of full waveforms. Powerful localization algorithms and automatic onset picking techniques (based on Akaikes Information Criterion) were established to handle the huge amount of data. Up to several thousand events were recorded during experiments of a few minutes. More sophisticated techniques like moment tensor inversion have been established on this relatively small scale as well. Problems are related to the amount of data but also to signal-to-noise quality, boundary conditions (reflections) sensor characteristics and unknown and changing Greens functions of the media. Some of the acoustic emissions recorded during these experiments had been transferred into audio range. The transformation into the audio range was done using Matlab. It is the aim of the sonification to establish a tool that is on one hand able to help controlling the experiment in-situ and probably adjust the load parameters according to the number and intensity of the acoustic emissions. On the other hand sonification can help to improve the understanding of acoustic emission techniques for training purposes (students, co-workers). On goal is to establish a real-time frequency transformation into the audio range to avoid time consuming visual data processing during the experiments. It is also the intention to analyze the signals using psycho-acoustic methods with the help of specialists from electrical engineering. Reference: Raith, Manuel (2013). "Schallemissionsanalyse bei Pulloutexperimenten an Verbunddübeln" Masterarbeit. Technische Universität München, Lehrstuhl für Zerstörungsfreie Prüfung. Malm, Fabian (2012). "Schallemissionsanalyse am humanen Femur" Masterarbeit. Technische Universität München, Lehrstuhl für Zerstörungsfreie Prüfung. Richter R. (2009): Einsatz der Schallemissionsanalyse zur Detektion des Riss und Abplatzungsverhaltens von Beton unter Brandeinwirkung. Diplomarbeit. Materialprüfungsanstalt Universität Stuttgart Keywords: Acoustic emission, bonded anchors, femur, pullout test, fire-spalling
Musician enhancement for speech-in-noise.
Parbery-Clark, Alexandra; Skoe, Erika; Lam, Carrie; Kraus, Nina
2009-12-01
To investigate the effect of musical training on speech-in-noise (SIN) performance, a complex task requiring the integration of working memory and stream segregation as well as the detection of time-varying perceptual cues. Previous research has indicated that, in combination with lifelong experience with musical stream segregation, musicians have better auditory perceptual skills and working memory. It was hypothesized that musicians would benefit from these factors and perform better on speech perception in noise than age-matched nonmusician controls. The performance of 16 musicians and 15 nonmusicians was compared on clinical measures of speech perception in noise-QuickSIN and Hearing-In-Noise Test (HINT). Working memory capacity and frequency discrimination were also assessed. All participants had normal hearing and were between the ages of 19 and 31 yr. To be categorized as a musician, participants needed to have started musical training before the age of 7 yr, have 10 or more years of consistent musical experience, and have practiced more than three times weekly within the 3 yr before study enrollment. Nonmusicians were categorized by the failure to meet the musician criteria, along with not having received musical training within the 7 yr before the study. Musicians outperformed the nonmusicians on both QuickSIN and HINT, in addition to having more fine-grained frequency discrimination and better working memory. Years of consistent musical practice correlated positively with QuickSIN, working memory, and frequency discrimination but not HINT. The results also indicate that working memory and frequency discrimination are more important for QuickSIN than for HINT. Musical experience appears to enhance the ability to hear speech in challenging listening environments. Large group differences were found for QuickSIN, and the results also suggest that this enhancement is derived in part from musicians' enhanced working memory and frequency discrimination. For HINT, in which performance was not linked to frequency discrimination ability and was only moderately linked to working memory, musicians still performed significantly better than the nonmusicians. The group differences for HINT were evident in the most difficult condition in which the speech and noise were presented from the same location and not spatially segregated. Understanding which cognitive and psychoacoustic factors as well as which lifelong experiences contribute to SIN may lead to more effective remediation programs for clinical populations for whom SIN poses a particular perceptual challenge. These results provide further evidence for musical training transferring to nonmusical domains and highlight the importance of taking musical training into consideration when evaluating a person's SIN ability in a clinical setting.
Validation of a Mobile Device for Acoustic Coordinated Reset Neuromodulation Tinnitus Therapy.
Hauptmann, Christian; Wegener, Alexander; Poppe, Hendrik; Williams, Mark; Popelka, Gerald; Tass, Peter A
2016-10-01
Sound-based tinnitus intervention stimuli include broad-band noise signals with subjectively adjusted bandwidths used as maskers delivered by commercial devices or hearing aids, environmental sounds broadly described and delivered by both consumer devices and hearing aids, music recordings specifically modified and delivered in a variety of different ways, and other stimuli. Acoustic coordinated reset neuromodulation therapy for tinnitus reduction has unique and more stringent requirements compared to all other sound-based tinnitus interventions. These include precise characterization of tinnitus pitch and loudness, and effective provision of patient-controlled daily therapy signals at defined frequencies, levels, and durations outside of the clinic. The purpose of this study was to evaluate an approach to accommodate these requirements including evaluation of a mobile device, validation of an automated tinnitus pitch-matching algorithm and assessment of a patient's ability to control stimuli and collect repeated outcome measures. The experimental design involved direct laboratory measurements of the sound delivery capabilities of a mobile device, comparison of an automated, adaptive pitch-matching method to a traditional manual method and measures of a patient's ability to understand and manipulate a mobile device graphic user interface to both deliver the therapy signals and collect the outcome measures. This study consisted of 5 samples of a common mobile device for the laboratory measures and a total of 30 adult participants: 15 randomly selected normal-hearing participants with simulated tinnitus for validation of a tinnitus pitch-matching algorithm and 15 sequentially selected patients already undergoing tinnitus therapy for evaluation of patient usability. No tinnitus intervention(s) were specifically studied as a component of this study. Data collection involved laboratory measures of mobile devices, comparison of manual and automated adaptive tinnitus pitch-matching psychoacoustic procedures in the same participant analyzed for absolute differences (t test), variance differences (f test), and range comparisons, and assessment of patient usability including questionnaire measures and logs of patient observations. Mobile devices are able to reliably and accurately deliver the acoustic therapy signals. There was no difference in mean pitch matches (t test, p > 0.05) between an automated adaptive method compared to a traditional manual pitch-matching method. However, the variability of the automated pitch-matching method was much less (f test, p < 0.05) with twice as many matches within the predefined error range (±5%) compared to the manual pitch-matching method (80% versus 40%). After a short initial training, all participants were able to use the mobile device effectively and to perform the required tasks without further professional assistance. American Academy of Audiology
Towards an Active Hearing Protection Device for Musicians =
NASA Astrophysics Data System (ADS)
Bernier, Antoine
Professional musicians are oftentimes exposed to high levels of sound. Prolonged or severe exposure to high sound levels could lead to permanent hearing loss and compromise their career. The logical solution would be to wear hearing protection devices (HPDs) when appropriate. However, perceptual discomfort associated with wearing HPD can discourage their use by musicians. The perceptual discomfort is caused by two detrimental effects: the occlusion effect and the isolation effect. The occlusion effect is often reported as an augmented, unnatural and annoying perception of one's own voice or instrument mechanically coupled to the head when wearing HPDs. The isolation effect is the unnatural sensation of being isolated from a given sound environment and can be caused by wearing HPDs that do not compensate for psychoacoustical factors and therefore alter the wearer's auditory perception. Both effects are highly unfavorable to the musicians' auditory perception and compromise their capacity to perform to the best of their abilities for their audience. They are among the reasons most often reported by musicians to decide not to wear HPDs. This master's project presents the concept and first prototype of an active HPD for musicians that aims at solving the detrimental effects while protecting the musician's hearing. A solution for the occlusion effect is presented in the form of an earplug complemented with in-ear active noise control. Practical design issues and required trade-off are analyzed through a literature review and the implementation and characterization of an active occlusion effect reduction system, allowing reduction of the occlusion effect between 8.5 and 12 dB at 250 Hz. A solution for the isolation effect is presented in the form of an earplug complemented with digital signal processing capabilities. Factors that may cause the isolation effect are identified through a literature review and corresponding algorithms that aim at re-establishing the naturalness of the auditory perception while wearing HPDs are presented through the design and implementation of an isolation effect compensation system, allowing up to 15 dB of variable uniform attenuation when used by itself. Both systems working simultaneously in the same device would result in an active HPD for musicians that reduces the occlusion effect and offers uniform variable attenuation up to 25 dB and perceived uniform attenuation up to 25 phons. The aim of this active HPD for musicians is to cause the least perceptual discomfort while protecting a musician's most precious tool: his hearing.
Bhatt, Ishan Sunilkumar
Tinnitus is a common otological condition that affects almost 10% of US adults. Research suggests that college students are vulnerable to tinnitus and hearing loss as they are exposed to traumatic levels of noise on a regular basis. Tinnitus and its influence in daily living continue to be underappreciated in the college-aged population. Therefore, the objective for the present study was to analyze prevalence and associated risk factors of tinnitus and tinnitus-related handicap in a sample of college-aged students. A survey was administered to 678 students aged 18-30 years in a cross-section of randomly selected university classes. The survey was adopted from the National Health and Nutrition Examination Survey (2010). It inquired about demographic details, medical and audiological history, routine noise exposure, smoking, sound level tolerance, tinnitus, and tinnitus-related handicap in daily living. Tinnitus-related handicap was assessed by the Tinnitus Handicap Inventory (THI). Participants were divided into four groups: chronic tinnitus (bothersome tinnitus for >1 year), acute tinnitus (bothersome tinnitus for ≤1 year), subacute tinnitus (at least one experience of tinnitus in a lifetime), and no tinnitus (no experience of tinnitus in a lifetime). The prevalence of chronic, acute, subacute, and no tinnitus was 8.4%, 13.0%, 37.9%, and 40.7% respectively. Almost 9% of subjects with any form of tinnitus reported more than a slight tinnitus-related handicap (i.e., THI score ≥18). A multinomial regression analysis revealed that individuals with high noise exposure, high sound level tolerance score, recurring ear infections, and self-reported hearing loss had high odds of chronic tinnitus. Females showed higher prevalence of acute tinnitus than males. Individuals with European American ethnicity and smoking history showed high odds of reporting subacute tinnitus. Almost 10% of the subjects reported that they were music students. The prevalence of chronic, acute, and subacute tinnitus was 11.3%, 22.5%, and 32.4%, respectively, for musicians, which was significantly higher than that for nonmusicians. Music exposure, firearm noise exposure, and occupational noise exposure were significantly correlated with tinnitus. Temporal characteristics of tinnitus, self-reported tinnitus loudness, and sound level tolerance were identified as major predictors for the overall THI score. Despite the reluctance to complain about tinnitus, a substantial portion of college-aged individuals reported tinnitus experience and its adverse influence in daily living. It was concluded that environmental and health-related factors can trigger tinnitus perception, while self-reported psychoacoustic descriptors of tinnitus may explain perceived tinnitus-related handicap in daily living by college-aged individuals. Future research is required to explore effects of tinnitus on educational achievements, social interaction, and vocational aspects of college students.
Firszt, Jill B; Reeder, Ruth M; Holden, Laura K
At a minimum, unilateral hearing loss (UHL) impairs sound localization ability and understanding speech in noisy environments, particularly if the loss is severe to profound. Accompanying the numerous negative consequences of UHL is considerable unexplained individual variability in the magnitude of its effects. Identification of covariables that affect outcome and contribute to variability in UHLs could augment counseling, treatment options, and rehabilitation. Cochlear implantation as a treatment for UHL is on the rise yet little is known about factors that could impact performance or whether there is a group at risk for poor cochlear implant outcomes when hearing is near-normal in one ear. The overall goal of our research is to investigate the range and source of variability in speech recognition in noise and localization among individuals with severe to profound UHL and thereby help determine factors relevant to decisions regarding cochlear implantation in this population. The present study evaluated adults with severe to profound UHL and adults with bilateral normal hearing. Measures included adaptive sentence understanding in diffuse restaurant noise, localization, roving-source speech recognition (words from 1 of 15 speakers in a 140° arc), and an adaptive speech-reception threshold psychoacoustic task with varied noise types and noise-source locations. There were three age-sex-matched groups: UHL (severe to profound hearing loss in one ear and normal hearing in the contralateral ear), normal hearing listening bilaterally, and normal hearing listening unilaterally. Although the normal-hearing-bilateral group scored significantly better and had less performance variability than UHLs on all measures, some UHL participants scored within the range of the normal-hearing-bilateral group on all measures. The normal-hearing participants listening unilaterally had better monosyllabic word understanding than UHLs for words presented on the blocked/deaf side but not the open/hearing side. In contrast, UHLs localized better than the normal-hearing unilateral listeners for stimuli on the open/hearing side but not the blocked/deaf side. This suggests that UHLs had learned strategies for improved localization on the side of the intact ear. The UHL and unilateral normal-hearing participant groups were not significantly different for speech in noise measures. UHL participants with childhood rather than recent hearing loss onset localized significantly better; however, these two groups did not differ for speech recognition in noise. Age at onset in UHL adults appears to affect localization ability differently than understanding speech in noise. Hearing thresholds were significantly correlated with speech recognition for UHL participants but not the other two groups. Auditory abilities of UHLs varied widely and could be explained only in part by hearing threshold levels. Age at onset and length of hearing loss influenced performance on some, but not all measures. Results support the need for a revised and diverse set of clinical measures, including sound localization, understanding speech in varied environments, and careful consideration of functional abilities as individuals with severe to profound UHL are being considered potential cochlear implant candidates.
Analysis of environmental sounds
NASA Astrophysics Data System (ADS)
Lee, Keansub
Environmental sound archives - casual recordings of people's daily life - are easily collected by MPS players or camcorders with low cost and high reliability, and shared in the web-sites. There are two kinds of user generated recordings we would like to be able to handle in this thesis: Continuous long-duration personal audio and Soundtracks of short consumer video clips. These environmental recordings contain a lot of useful information (semantic concepts) related with activity, location, occasion and content. As a consequence, the environment archives present many new opportunities for the automatic extraction of information that can be used in intelligent browsing systems. This thesis proposes systems for detecting these interesting concepts on a collection of these real-world recordings. The first system is to segment and label personal audio archives - continuous recordings of an individual's everyday experiences - into 'episodes' (relatively consistent acoustic situations lasting a few minutes or more) using the Bayesian Information Criterion and spectral clustering. The second system is for identifying regions of speech or music in the kinds of energetic and highly-variable noise present in this real-world sound. Motivated by psychoacoustic evidence that pitch is crucial in the perception and organization of sound, we develop a noise-robust pitch detection algorithm to locate speech or music-like regions. To avoid false alarms resulting from background noise with strong periodic components (such as air-conditioning), a new scheme is added in order to suppress these noises in the domain of autocorrelogram. In addition, the third system is to automatically detect a large set of interesting semantic concepts; which we chose for being both informative and useful to users, as well as being technically feasible. These 25 concepts are associated with people's activities, locations, occasions, objects, scenes and sounds, and are based on a large collection of consumer videos in conjunction with user studies. We model the soundtrack of each video, regardless of its original duration, as a fixed-sized clip-level summary feature. For each concept, an SVM-based classifier is trained according to three distance measures (Kullback-Leibler, Bhattacharyya, and Mahalanobis distance). Detecting the time of occurrence of a local object (for instance, a cheering sound) embedded in a longer soundtrack is useful and important for applications such as search and retrieval in consumer video archives. We finally present a Markov-model based clustering algorithm able to identify and segment consistent sets of temporal frames into regions associated with different ground-truth labels, and at the same time to exclude a set of uninformative frames shared in common from all clips. The labels are provided at the clip level, so this refinement of the time axis represents a variant of Multiple-Instance Learning (MIL). Quantitative evaluation shows that the performance of our proposed approaches tested on the 60h personal audio archives or 1900 YouTube video clips is significantly better than existing algorithms for detecting these useful concepts in real-world personal audio recordings.
Bernstein, Joshua G.W.; Mehraei, Golbarg; Shamma, Shihab; Gallun, Frederick J.; Theodoroff, Sarah M.; Leek, Marjorie R.
2014-01-01
Background A model that can accurately predict speech intelligibility for a given hearing-impaired (HI) listener would be an important tool for hearing-aid fitting or hearing-aid algorithm development. Existing speech-intelligibility models do not incorporate variability in suprathreshold deficits that are not well predicted by classical audiometric measures. One possible approach to the incorporation of such deficits is to base intelligibility predictions on sensitivity to simultaneously spectrally and temporally modulated signals. Purpose The likelihood of success of this approach was evaluated by comparing estimates of spectrotemporal modulation (STM) sensitivity to speech intelligibility and to psychoacoustic estimates of frequency selectivity and temporal fine-structure (TFS) sensitivity across a group of HI listeners. Research Design The minimum modulation depth required to detect STM applied to an 86 dB SPL four-octave noise carrier was measured for combinations of temporal modulation rate (4, 12, or 32 Hz) and spectral modulation density (0.5, 1, 2, or 4 cycles/octave). STM sensitivity estimates for individual HI listeners were compared to estimates of frequency selectivity (measured using the notched-noise method at 500, 1000measured using the notched-noise method at 500, 2000, and 4000 Hz), TFS processing ability (2 Hz frequency-modulation detection thresholds for 500, 10002 Hz frequency-modulation detection thresholds for 500, 2000, and 4000 Hz carriers) and sentence intelligibility in noise (at a 0 dB signal-to-noise ratio) that were measured for the same listeners in a separate study. Study Sample Eight normal-hearing (NH) listeners and 12 listeners with a diagnosis of bilateral sensorineural hearing loss participated. Data Collection and Analysis STM sensitivity was compared between NH and HI listener groups using a repeated-measures analysis of variance. A stepwise regression analysis compared STM sensitivity for individual HI listeners to audiometric thresholds, age, and measures of frequency selectivity and TFS processing ability. A second stepwise regression analysis compared speech intelligibility to STM sensitivity and the audiogram-based Speech Intelligibility Index. Results STM detection thresholds were elevated for the HI listeners, but only for low rates and high densities. STM sensitivity for individual HI listeners was well predicted by a combination of estimates of frequency selectivity at 4000 Hz and TFS sensitivity at 500 Hz but was unrelated to audiometric thresholds. STM sensitivity accounted for an additional 40% of the variance in speech intelligibility beyond the 40% accounted for by the audibility-based Speech Intelligibility Index. Conclusions Impaired STM sensitivity likely results from a combination of a reduced ability to resolve spectral peaks and a reduced ability to use TFS information to follow spectral-peak movements. Combining STM sensitivity estimates with audiometric threshold measures for individual HI listeners provided a more accurate prediction of speech intelligibility than audiometric measures alone. These results suggest a significant likelihood of success for an STM-based model of speech intelligibility for HI listeners. PMID:23636210
Sidiras, Chris; Iliadou, Vasiliki Vivian; Chermak, Gail D; Nimatoudis, Ioannis
2016-05-01
Including speech recognition in noise testing in audiological evaluations may reveal functional hearing deficits that may otherwise remain undetected. The current study explored the potential utility of the Speech-in-Babble (SinB) test in the assessment of central auditory processing disorder (CAPD) in young children for whom diagnosis is challenging. A cross-sectional analysis. Forty-one Greek children 4-13 yr of age diagnosed with CAPD and exhibiting listening and academic problems (clinical group) and 20 age-matched controls with no listening or academic problems participated in the study. All participants' auditory processing was assessed using the same tests and instrumentation in a sound-treated room. Two equivalent lists of the SinB test, developed at the Psychoacoustic Laboratory of the Aristotle University of Thessaloniki, were administered monaurally in a counterbalanced order. SinB consists of lists of 50 phonetically balanced disyllabic words presented in background multitalker babble. Five signal-to-noise ratios (SNRs) were used in a fixed order. The children were instructed to repeat the word after each presentation. The SNR at which the child achieved 50% correct word identification served as the dependent variable or outcome measure, with higher SinB scores (measured in SNR dB) corresponding to poorer performance. SinB performance was better (lower SNR) for the normal control group versus the clinical group [F(1,35) = 43.03, p < 0.0001]. SinB inversely correlated with age for both CAPD and control groups (r = -0.648, p < 0.001 and r = -0.658, p < 0.005, respectively). Regression analysis revealed that linear models better explained the variance in the data than a quadratic model for both the control and CAPD groups. The slope (beta value of the linear model) was steeper for the clinical group compared to the control group (beta = -0.306 versus beta = -0.130, respectively). An analysis of covariance run with age as the covariate to assess the potential effect of comorbidity on SinB performance in children with CAPD with and without comorbid conditions revealed no significant differences between groups [F(1,38) = 0.149, p > 0.05]. This study offers the first detailed presentation of the performance of Greek children on a Greek language SinB test. The main finding is that SinB scores improved as a function of age in a constant manner as represented by the slope of the linear regression line for both CAPD and control groups. Results suggest that this speech recognition in competition test holds promise for differentiating typically developing Greek children from those children with CAPD across the age range studied here (4-13 yr). The SinB seemed rather immune to the presence of comorbid conditions presented by some of the children in this study, suggesting its potential utility as a valid measure of central auditory processing. While there are many speech-in-noise or competition tests in English, there are fewer in other languages. Tests like the SinB should be developed in other languages to ensure that children demonstrating "listening" problems can be properly evaluated. American Academy of Audiology.
Firszt, Jill B.; Reeder, Ruth M.; Holden, Laura K.
2016-01-01
Objectives At a minimum, unilateral hearing loss (UHL) impairs sound localization ability and understanding speech in noisy environments, particularly if the loss is severe to profound. Accompanying the numerous negative consequences of UHL is considerable unexplained individual variability in the magnitude of its effects. Identification of co-variables that affect outcome and contribute to variability in UHLs could augment counseling, treatment options, and rehabilitation. Cochlear implantation as a treatment for UHL is on the rise yet little is known about factors that could impact performance or whether there is a group at risk for poor cochlear implant outcomes when hearing is near-normal in one ear. The overall goal of our research is to investigate the range and source of variability in speech recognition in noise and localization among individuals with severe to profound UHL and thereby help determine factors relevant to decisions regarding cochlear implantation in this population. Design The present study evaluated adults with severe to profound UHL and adults with bilateral normal hearing. Measures included adaptive sentence understanding in diffuse restaurant noise, localization, roving-source speech recognition (words from 1 of 15 speakers in a 140° arc) and an adaptive speech-reception threshold psychoacoustic task with varied noise types and noise-source locations. There were three age-gender-matched groups: UHL (severe to profound hearing loss in one ear and normal hearing in the contralateral ear), normal hearing listening bilaterally, and normal hearing listening unilaterally. Results Although the normal-hearing-bilateral group scored significantly better and had less performance variability than UHLs on all measures, some UHL participants scored within the range of the normal-hearing-bilateral group on all measures. The normal-hearing participants listening unilaterally had better monosyllabic word understanding than UHLs for words presented on the blocked/deaf side but not the open/hearing side. In contrast, UHLs localized better than the normal hearing unilateral listeners for stimuli on the open/hearing side but not the blocked/deaf side. This suggests that UHLs had learned strategies for improved localization on the side of the intact ear. The UHL and unilateral normal hearing participant groups were not significantly different for speech-in-noise measures. UHL participants with childhood rather than recent hearing loss onset localized significantly better; however, these two groups did not differ for speech recognition in noise. Age at onset in UHL adults appears to affect localization ability differently than understanding speech in noise. Hearing thresholds were significantly correlated with speech recognition for UHL participants but not the other two groups. Conclusions Auditory abilities of UHLs varied widely and could be explained only in part by hearing threshold levels. Age at onset and length of hearing loss influenced performance on some, but not all measures. Results support the need for a revised and diverse set of clinical measures, including sound localization, understanding speech in varied environments and careful consideration of functional abilities as individuals with severe to profound UHL are being considered potential cochlear implant candidates. PMID:28067750
Woodwind Tone Hole Acoustics and the Spectrum Transformation Function.
NASA Astrophysics Data System (ADS)
Keefe, Douglas Howard
This report describes an investigation of woodwind musical instrument tone holes and their effect on the radiated spectrum, the total dissipation, the stability of oscillation, the psychoacoustical cues important in perception, and the tuning and response of the instrument. Varying tone hole proportions significantly affect the radiative and frictional damping near a single hole, the mutual interactions between holes, the onset of streaming and turbulence near the holes, and the perceived woodwind timbre. The interconnections between related fields are explored through a brief review of sound production in woodwinds plus more extensive reviews of room and psychological acoustics. A theoretical and experimental discussion of the spectrum transformation function from the mouthpiece into the room relates all these fields. Also, considered are differences between cylindrical and conical bore woodwinds, the systematic shifts in saxophone spectra produced by the beating of the reed, the coupling of many closely spaced tone holes to the room excitation, the role of the player, and the results pertaining to computer music synthesis. The complicated acoustical flow inside the main air column near a single tone hole has been examined using a Green function, integral equation approach. A variational formulation allows explicit calculation of the open and closed hole impedance parameters needed in the transmission line description of a woodwind, and experiments have verified the theory in detail. Major acoustical topics considered are listed below. The effective length t(,e) of an open hole, relevant for instrument design and modification, is calculated and measured in terms of the main bore diameter 2a, hole diameter 2b, and the height t of the hole chimney; the effect of a hanging pad is a semi-empirical correction on t(,e). When the fundamental plane-wave mode of the main air column oscillation is at a pressure node, both the open and closed hole series impedances are negative inertances whose values depend on the tone hole proportions. An open hole at a pressure node can radiate as a dipole when (b/a) is large and (t/2b) is small. Dissipative losses vary significantly with the frequency of oscillation and the tone hole geometry. Lowering the pad height above a tone hole increases the dissipation. Acoustical streaming through holes is very important for t << 2b, and the associated nonlinear dissipation can destroy the oscillation on poorly designed woodwinds. This unexpected phenomenon is critical in the playing behavior of some flutes, clarinets, and other woodwinds. The onset of streaming occurs at all dynamical levels and more easily for instruments whose spectra are in a 1:3:5 frequency ratio, rather than a 1:2:3 ratio. The streaming is most important for low register tones for which the usual dissipation is also the largest relative to the radiative dissipation, due to losses at the sharp edges inside the bore near the tone holes. Mutual interactions between holes separated by a distance 2s are most pronounced for large diameter holes (2b/2s not small). Holes interact externally via radiation, and internally via higher-order evanescent modes excited at the intersection of the main bore with each tone hole. The non-radiative dissipation increases, and the air column resonances are slightly shifted due to the presence of these interactions. Applications are discussed and numerous additional experiments are proposed which are relevant to woodwinds and their design, and the perception of listeners in rooms.