Kastelein, Ronald A; Wensveen, Paul J; Terhune, John M; de Jong, Christ A F
2011-01-01
Equal-loudness functions describe relationships between the frequencies of sounds and their perceived loudness. This pilot study investigated the possibility of deriving equal-loudness contours based on the assumption that sounds of equal perceived loudness elicit equal reaction times (RTs). During a psychoacoustic underwater hearing study, the responses of two young female harbor seals to tonal signals between 0.125 and 100 kHz were filmed. Frame-by-frame analysis was used to quantify RT (the time between the onset of the sound stimulus and the onset of movement of the seal away from the listening station). Near-threshold equal-latency contours, as surrogates for equal-loudness contours, were estimated from RT-level functions fitted to mean RT data. The closer the received sound pressure level was to the 50% detection hearing threshold, the more slowly the animals reacted to the signal (RT range: 188-982 ms). Equal-latency contours were calculated relative to the RTs shown by each seal at sound levels of 0, 10, and 20 dB above the detection threshold at 1 kHz. Fifty percent detection thresholds are obtained with well-trained subjects actively listening for faint familiar sounds. When calculating audibility ranges of sounds for harbor seals in nature, it may be appropriate to consider levels 20 dB above this threshold.
Wensveen, Paul J; Huijser, Léonie A E; Hoek, Lean; Kastelein, Ronald A
2016-01-01
Loudness perception can be studied based on the assumption that sounds of equal loudness elicit equal reaction time (RT; or "response latency"). We measured the underwater RTs of a harbor porpoise to narrowband frequency-modulated sounds and constructed six equal-latency contours. The contours paralleled the audiogram at low sensation levels (high RTs). At high-sensation levels, contours flattened between 0.5 and 31.5 kHz but dropped substantially (RTs shortened) beyond those frequencies. This study suggests that equal-latency-based frequency weighting can emulate noise perception in porpoises for low and middle frequencies but that the RT-loudness correlation is relatively weak for very high frequencies.
Visualization of Sound Waves Using Regularly Spaced Soap Films
ERIC Educational Resources Information Center
Elias, F.; Hutzler, S.; Ferreira, M. S.
2007-01-01
We describe a novel demonstration experiment for the visualization and measurement of standing sound waves in a tube. The tube is filled with equally spaced soap films whose thickness varies in response to the amplitude of the sound wave. The thickness variations are made visible based on optical interference. The distance between two antinodes is…
Kuroda, Tsuyoshi; Tomimatsu, Erika; Grondin, Simon; Miyazaki, Makoto
2016-11-01
We investigated how perceived duration of empty time intervals would be modulated by the length of sounds marking those intervals. Three sounds were successively presented in Experiment 1. Each sound was short (S) or long (L), and the temporal position of the middle sound's onset was varied. The lengthening of each sound resulted in delayed perception of the onset; thus, the middle sound's onset had to be presented earlier in the SLS than in the LSL sequence so that participants perceived the three sounds as presented at equal interonset intervals. In Experiment 2, a short sound and a long sound were alternated repeatedly, and the relative duration of the SL interval to the LS interval was varied. This repeated sequence was perceived as consisting of equal interonset intervals when the onsets of all sounds were aligned at physically equal intervals. If the same onset delay as in the preceding experiment had occurred, participants should have perceived equality between the interonset intervals in the repeated sequence when the SL interval was physically shortened relative to the LS interval. The effects of sound length seemed to be canceled out when the presentation of intervals was repeated. Finally, the perceived duration of the interonset intervals in the repeated sequence was not influenced by whether the participant's native language was French or Japanese, or by how the repeated sequence was perceptually segmented into rhythmic groups.
Perception of Water-Based Masking Sounds-Long-Term Experiment in an Open-Plan Office.
Hongisto, Valtteri; Varjo, Johanna; Oliva, David; Haapakangas, Annu; Benway, Evan
2017-01-01
A certain level of masking sound is necessary to control the disturbance caused by speech sounds in open-plan offices. The sound is usually provided with evenly distributed loudspeakers. Pseudo-random noise is often used as a source of artificial sound masking (PRMS). A recent laboratory experiment suggested that water-based masking sound (WBMS) could be more favorable than PRMS. The purpose of our study was to determine how the employees perceived different WBMSs compared to PRMS. The experiment was conducted in an open-plan office of 77 employees who had been accustomed to work under PRMS (44 dB L Aeq ). The experiment consisted of five masking conditions: the original PRMS, four different WBMSs and return to the original PRMS. The exposure time of each condition was 3 weeks. The noise level was nearly equal between the conditions (43-45 dB L Aeq ) but the spectra and the nature of the sounds were very different. A questionnaire was completed at the end of each condition. Acoustic satisfaction was worse during the WBMSs than during the PRMS. The disturbance caused by three out of four WBMSs was larger than that of PRMS. Several attributes describing the sound quality itself were in favor of PRMS. Colleagues' speech sounds disturbed more during WBMSs. None of the WBMSs produced better subjective ratings than PRMS. Although the first WBMS was equal with the PRMS for several variables, the overall results cannot be seen to support the use of WBMSs in office workplaces. Because the experiment suffered from some methodological weaknesses, conclusions about the adequacy of WBMSs cannot yet be drawn.
Category 5: Sound Generation In Viscous Problems. Problem 2: Sound Generation By Flow Over a Cavity
NASA Technical Reports Server (NTRS)
Henderson, Brenda S.
2004-01-01
The discrete frequency sound produced by the flow of air at low subsonic speeds over a deep cavity was investigated. A long aspect ratio rectangular cavity with a leading edge overhang that cut off of the cavity opening was placed flush with the top surface of a wind tunnel. The approach flow velocity was maintained at 50 m/s for the benchmark problem although results are also presented for other conditions. Boundary layer measurements conducted with a single element hotwire anemometer indicated that the boundary layer thickness just upstream of the cavity was equal to 17 mm. Sound pressure level measurements were made at three locations in the cavity: the center of the leading edge wall, the center of the cavity floor, and the center of the trailing edge wall. Three discrete tones were measured at all three locations with corresponding Strouhal numbers (based on cavity opening length and approach flow velocity) equal to 0.24, 0.26, and 0.41. The amplitudes of each tone were approximately equal at each measurement location in the cavity. Measurements made at other approach flow conditions indicated that the approach flow velocity and the boundary layer thickness affected the frequency characteristics of the discrete tones.
Perception of Water-Based Masking Sounds—Long-Term Experiment in an Open-Plan Office
Hongisto, Valtteri; Varjo, Johanna; Oliva, David; Haapakangas, Annu; Benway, Evan
2017-01-01
A certain level of masking sound is necessary to control the disturbance caused by speech sounds in open-plan offices. The sound is usually provided with evenly distributed loudspeakers. Pseudo-random noise is often used as a source of artificial sound masking (PRMS). A recent laboratory experiment suggested that water-based masking sound (WBMS) could be more favorable than PRMS. The purpose of our study was to determine how the employees perceived different WBMSs compared to PRMS. The experiment was conducted in an open-plan office of 77 employees who had been accustomed to work under PRMS (44 dB LAeq). The experiment consisted of five masking conditions: the original PRMS, four different WBMSs and return to the original PRMS. The exposure time of each condition was 3 weeks. The noise level was nearly equal between the conditions (43–45 dB LAeq) but the spectra and the nature of the sounds were very different. A questionnaire was completed at the end of each condition. Acoustic satisfaction was worse during the WBMSs than during the PRMS. The disturbance caused by three out of four WBMSs was larger than that of PRMS. Several attributes describing the sound quality itself were in favor of PRMS. Colleagues' speech sounds disturbed more during WBMSs. None of the WBMSs produced better subjective ratings than PRMS. Although the first WBMS was equal with the PRMS for several variables, the overall results cannot be seen to support the use of WBMSs in office workplaces. Because the experiment suffered from some methodological weaknesses, conclusions about the adequacy of WBMSs cannot yet be drawn. PMID:28769834
NASA Astrophysics Data System (ADS)
Gauthier, P.-A.; Camier, C.; Lebel, F.-A.; Pasco, Y.; Berry, A.; Langlois, J.; Verron, C.; Guastavino, C.
2016-08-01
Sound environment reproduction of various flight conditions in aircraft mock-ups is a valuable tool for the study, prediction, demonstration and jury testing of interior aircraft sound quality and annoyance. To provide a faithful reproduced sound environment, time, frequency and spatial characteristics should be preserved. Physical sound field reproduction methods for spatial sound reproduction are mandatory to immerse the listener's body in the proper sound fields so that localization cues are recreated at the listener's ears. Vehicle mock-ups pose specific problems for sound field reproduction. Confined spaces, needs for invisible sound sources and very specific acoustical environment make the use of open-loop sound field reproduction technologies such as wave field synthesis (based on free-field models of monopole sources) not ideal. In this paper, experiments in an aircraft mock-up with multichannel least-square methods and equalization are reported. The novelty is the actual implementation of sound field reproduction with 3180 transfer paths and trim panel reproduction sources in laboratory conditions with a synthetic target sound field. The paper presents objective evaluations of reproduced sound fields using various metrics as well as sound field extrapolation and sound field characterization.
Laboratory studies of scales for measuring helicopter noise
NASA Technical Reports Server (NTRS)
Ollerhead, J. B.
1982-01-01
The adequacy of the effective perceived noise level (EPNL) procedure for rating helicopter noise annoyance was investigated. Recordings of 89 helicopters and 30 fixed wing aircraft (CTOL) flyover sounds were rated with respect to annoyance by groups of approximately 40 subjects. The average annoyance scores were transformed to annoyance levels defined as the equally annoying sound levels of a fixed reference sound. The sound levels of the test sounds were measured on various scales, with and without corrections for duration, tones, and impulsiveness. On average, the helicopter sounds were judged equally annoying to CTOL sounds when their duration corrected levels are approximately 2 dB higher. Multiple regression analysis indicated that, provided the helicopter/CTOL difference of about 2 dB is taken into account, the particular linear combination of level, duration, and tone corrections inherent in EPNL is close to optimum. The results reveal no general requirement for special EPNL correction terms to penalize helicopter sounds which are particularly impulsive; impulsiveness causes spectral and temporal changes which themselves adequately amplify conventionally measured sound levels.
Sound Propagation in Shallow Water with an Inhomogeneous GAS-Saturated Bottom
NASA Astrophysics Data System (ADS)
Grigor'ev, V. A.; Petnikov, V. G.; Roslyakov, A. G.; Terekhina, Ya. E.
2018-05-01
We present the methods and results of numerical experiments studying the low-frequency sound propagation in one of the areas of the Arctic shelf with a randomly inhomogeneous gas-saturated bottom. The characteristics of the upper layer of bottom sedimentary rocks (sediments) used in calculations were obtained during a 3D seismic survey and trial drilling of the seafloor. We demonstrate the possibilities of substituting in numerical simulation a real bottom with a fluid homogeneous half-space where the effective value of the sound speed is equal to the average sound speed in the bottom, with averaging along the sound propagation path to a sediment depth of 0.6 wavelength in the bottom. An original technique is proposed for estimating the sound speed propagation in an upper inhomogeneous sediment layer. The technique is based on measurements of acoustic wave attenuation in water during waveguide propagation.
Acoustic Performance of a Real-Time Three-Dimensional Sound-Reproduction System
NASA Technical Reports Server (NTRS)
Faller, Kenneth J., II; Rizzi, Stephen A.; Aumann, Aric R.
2013-01-01
The Exterior Effects Room (EER) is a 39-seat auditorium at the NASA Langley Research Center and was built to support psychoacoustic studies of aircraft community noise. The EER has a real-time simulation environment which includes a three-dimensional sound-reproduction system. This system requires real-time application of equalization filters to compensate for spectral coloration of the sound reproduction due to installation and room effects. This paper describes the efforts taken to develop the equalization filters for use in the real-time sound-reproduction system and the subsequent analysis of the system s acoustic performance. The acoustic performance of the compensated and uncompensated sound-reproduction system is assessed for its crossover performance, its performance under stationary and dynamic conditions, the maximum spatialized sound pressure level it can produce from a single virtual source, and for the spatial uniformity of a generated sound field. Additionally, application examples are given to illustrate the compensated sound-reproduction system performance using recorded aircraft flyovers
Towards a unifying basis of auditory thresholds: binaural summation.
Heil, Peter
2014-04-01
Absolute auditory threshold decreases with increasing sound duration, a phenomenon explainable by the assumptions that the sound evokes neural events whose probabilities of occurrence are proportional to the sound's amplitude raised to an exponent of about 3 and that a constant number of events are required for threshold (Heil and Neubauer, Proc Natl Acad Sci USA 100:6151-6156, 2003). Based on this probabilistic model and on the assumption of perfect binaural summation, an equation is derived here that provides an explicit expression of the binaural threshold as a function of the two monaural thresholds, irrespective of whether they are equal or unequal, and of the exponent in the model. For exponents >0, the predicted binaural advantage is largest when the two monaural thresholds are equal and decreases towards zero as the monaural threshold difference increases. This equation is tested and the exponent derived by comparing binaural thresholds with those predicted on the basis of the two monaural thresholds for different values of the exponent. The thresholds, measured in a large sample of human subjects with equal and unequal monaural thresholds and for stimuli with different temporal envelopes, are compatible only with an exponent close to 3. An exponent of 3 predicts a binaural advantage of 2 dB when the two ears are equally sensitive. Thus, listening with two (equally sensitive) ears rather than one has the same effect on absolute threshold as doubling duration. The data suggest that perfect binaural summation occurs at threshold and that peripheral neural signals are governed by an exponent close to 3. They might also shed new light on mechanisms underlying binaural summation of loudness.
Visual sign phonology: insights into human reading and language from a natural soundless phonology.
Petitto, L A; Langdon, C; Stone, A; Andriola, D; Kartheiser, G; Cochran, C
2016-11-01
Among the most prevailing assumptions in science and society about the human reading process is that sound and sound-based phonology are critical to young readers. The child's sound-to-letter decoding is viewed as universal and vital to deriving meaning from print. We offer a different view. The crucial link for early reading success is not between segmental sounds and print. Instead the human brain's capacity to segment, categorize, and discern linguistic patterning makes possible the capacity to segment all languages. This biological process includes the segmentation of languages on the hands in signed languages. Exposure to natural sign language in early life equally affords the child's discovery of silent segmental units in visual sign phonology (VSP) that can also facilitate segmental decoding of print. We consider powerful biological evidence about the brain, how it builds sound and sign phonology, and why sound and sign phonology are equally important in language learning and reading. We offer a testable theoretical account, reading model, and predictions about how VSP can facilitate segmentation and mapping between print and meaning. We explain how VSP can be a powerful facilitator of all children's reading success (deaf and hearing)-an account with profound transformative impact on learning to read in deaf children with different language backgrounds. The existence of VSP has important implications for understanding core properties of all human language and reading, challenges assumptions about language and reading as being tied to sound, and provides novel insight into a remarkable biological equivalence in signed and spoken languages. WIREs Cogn Sci 2016, 7:366-381. doi: 10.1002/wcs.1404 For further resources related to this article, please visit the WIREs website. © 2016 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Zhao, Liang; Ge, Jian-Hua
2012-12-01
Single-carrier (SC) transmission with frequency-domain equalization (FDE) is today recognized as an attractive alternative to orthogonal frequency-division multiplexing (OFDM) for communication application with the inter-symbol interference (ISI) caused by multi-path propagation, especially in shallow water channel. In this paper, we investigate an iterative receiver based on minimum mean square error (MMSE) decision feedback equalizer (DFE) with symbol rate and fractional rate samplings in the frequency domain (FD) and serially concatenated trellis coded modulation (SCTCM) decoder. Based on sound speed profiles (SSP) measured in the lake and finite-element ray tracking (Bellhop) method, the shallow water channel is constructed to evaluate the performance of the proposed iterative receiver. Performance results show that the proposed iterative receiver can significantly improve the performance and obtain better data transmission than FD linear and adaptive decision feedback equalizers, especially in adopting fractional rate sampling.
Educational Equality: Luck Egalitarian, Pluralist and Complex
ERIC Educational Resources Information Center
Calvert, John
2014-01-01
The basic principle of educational equality is that each child should receive an equally good education. This sounds appealing, but is rather vague and needs substantial working out. Also, educational equality faces all the objections to equality per se, plus others specific to its subject matter. Together these have eroded confidence in the…
Are Masking-Based Models of Risk Useful?
Gisiner, Robert C
2016-01-01
As our understanding of directly observable effects from anthropogenic sound exposure has improved, concern about "unobservable" effects such as stress and masking have received greater attention. Equal energy models of masking such as power spectrum models have the appeal of simplicity, but do they offer biologically realistic assessments of the risk of masking? Data relevant to masking such as critical ratios, critical bandwidths, temporal resolution, and directional resolution along with what is known about general mammalian antimasking mechanisms all argue for a much more complicated view of masking when making decisions about the risk of masking inherent in a given anthropogenic sound exposure scenario.
Sound Effects for Children's Comprehension of Variably-Paced Television Programs.
ERIC Educational Resources Information Center
Calvert, Sandra L.; Scott, M. Catherine
In this study, children's selective attention to, and comprehension of, variably-paced television programs were examined as a function of sound effects. Sixty-four children, equally distributed by sex and by preschool and fourth grades, were randomly assigned to one of four treatment conditions which crossed two levels of sound effects (presence…
Loudness of steady sounds - A new theory
NASA Technical Reports Server (NTRS)
Howes, W. L.
1979-01-01
A new mathematical theory for calculating the loudness of steady sounds from power summation and frequency interaction, based on psychoacoustic and physiological information, assuems that loudness is a subjective measure of the electrical energy transmitted along the auditory nerve to the central nervous system. The auditory system consists of the mechanical part modeled by a bandpass filter with a transfer function dependent on the sound pressure, and the electrical part where the signal is transformed into a half-wave reproduction represented by the electrical power in impulsive discharges transmitted along neurons comprising the auditory nerve. In the electrical part the neurons are distributed among artificial parallel channels with frequency bandwidths equal to 'critical bandwidths for loudness', within which loudness is constant for constant sound pressure. The total energy transmitted to the central nervous system is the sum of the energy transmitted in all channels, and the loudness is proportional to the square root of the total filtered sound energy distributed over all channels. The theory explains many psychoacoustic phenomena such as audible beats resulting from closely spaced tones, interaction of sound stimuli which affect the same neurons affecting loudness, and of individually subliminal sounds becoming audible if they lie within the same critical band.
NASA Technical Reports Server (NTRS)
Brown, David; Sutherland, Louis C.
1992-01-01
The preferred descriptor to define the spectral content of sonic booms is the Sound Exposure Spectrum Level, LE(f). This descriptor represents the spectral content of the basic noise descriptors used for describing any single event--the Sound Exposure Level, LE. The latter is equal to ten times the logarithms, to the base ten, of the integral, over the duration of the event, of the square of the instantaneous acoustic pressure, divided by the square of the reference pressure, 20 micro-Pa. When applied to the evaluation of community response to sonic booms, it is customary to use the so-called C-Weighted Sound Exposure Level, LCE, for which the frequency content of the instantaneous acoustic pressure is modified by the C-Weighting curve.
Musician's and physicist's view on tuning keyboard instruments
NASA Astrophysics Data System (ADS)
Lubenow, Martin; Meyn, Jan-Peter
2007-01-01
The simultaneous sound of several voices or instruments requires proper tuning to achieve consonance for certain intervals and chords. Most instruments allow enough frequency variation to enable pure tuning while being played. Keyboard instruments such as organ and piano have given frequencies for individual notes and the tuning must be based on a compromise. The equal temperament is not the only solution, but a special choice. Unequal temperaments produce better results in many cases, because important major thirds and triads are improved. Equal temperament was not propagated by Johann Sebastian Bach, as is often stated in introductory literature on this topic.
On the diffusion of sound in an auditorium
NASA Astrophysics Data System (ADS)
Harris, Cyril M.
2005-09-01
A condition of perfect diffusion of sound is said to exist in an auditorium if, at any point within it, the reverberant sound travels in all directions with equal probability, and if the level of the reflected sound is everywhere equal. In deriving the reverberation time formula, which predicts how long sound will bounce around an enclosed space after the source has stopped, W.C. Sabine assumed perfect diffusion within it. When this is not the case, his formula may predict inaccurate results. For example, the Sabine equation will not give correct results in an auditorium with poor diffusion, as when there is a large overhanging balcony, or if one of the dimensions of the enclosed space is very much greater than the other dimensions, or if the auditorium is divided into spaces having different acoustical properties. An auditorium with excellent diffusion beneficially affects the uniformity of decay of sound within the space and pleases the listener's ear. Among techniques that contribute to good diffusion are the surface irregularities found in the elaborate styles of architecture of the past. Illustrations will be presented showing some approaches within the modern architectural idiom that have yielded successful results.
Popov, Vladimir V; Supin, Alexander Ya; Rozhnov, Viatcheslav V; Nechaev, Dmitry I; Sysueva, Evgenia V
2014-05-15
The influence of fatiguing sound level and duration on post-exposure temporary threshold shift (TTS) was investigated in two beluga whales (Delphinapterus leucas). The fatiguing sound was half-octave noise with a center frequency of 22.5 kHz. TTS was measured at a test frequency of 32 kHz. Thresholds were measured by recording rhythmic evoked potentials (the envelope following response) to a test series of short (eight cycles) tone pips with a pip rate of 1000 s(-1). TTS increased approximately proportionally to the dB measure of both sound pressure (sound pressure level, SPL) and duration of the fatiguing noise, as a product of these two variables. In particular, when the noise parameters varied in a manner that maintained the product of squared sound pressure and time (sound exposure level, SEL, which is equivalent to the overall noise energy) at a constant level, TTS was not constant. Keeping SEL constant, the highest TTS appeared at an intermediate ratio of SPL to sound duration and decreased at both higher and lower ratios. Multiplication (SPL multiplied by log duration) better described the experimental data than an equal-energy (equal SEL) model. The use of SEL as a sole universal metric may result in an implausible assessment of the impact of a fatiguing sound on hearing thresholds in odontocetes, including under-evaluation of potential risks. © 2014. Published by The Company of Biologists Ltd.
A new method for the automatic interpretation of Schlumberger and Wenner sounding curves
Zohdy, A.A.R.
1989-01-01
A fast iterative method for the automatic interpretation of Schlumberger and Wenner sounding curves is based on obtaining interpreted depths and resistivities from shifted electrode spacings and adjusted apparent resistivities, respectively. The method is fully automatic. It does not require an initial guess of the number of layers, their thicknesses, or their resistivities; and it does not require extrapolation of incomplete sounding curves. The number of layers in the interpreted model equals the number of digitized points on the sounding curve. The resulting multilayer model is always well-behaved with no thin layers of unusually high or unusually low resistivities. For noisy data, interpretation is done in two sets of iterations (two passes). Anomalous layers, created because of noise in the first pass, are eliminated in the second pass. Such layers are eliminated by considering the best-fitting curve from the first pass to be a smoothed version of the observed curve and automatically reinterpreting it (second pass). The application of the method is illustrated by several examples. -Author
The Great Lake Erie. A Reference Text for Educators and Communicators.
ERIC Educational Resources Information Center
Fortner, Rosanne W., Ed.; Mayer, Victor J., Ed.
The purpose of this volume is to present a body of basic information about the Great Lakes that is current and based on sound research. Such information is frequently difficult to locate, especially in one reference, and equally difficult to decipher and evaluate. The 16 essays contained in this book deal with a variety of topics including the…
NASA Astrophysics Data System (ADS)
Long, Derle Ray
Coincidence theory states that when the components of harmony are in enhanced alignment the sound will be more consonant to the human auditory system. An objective method of examining the components of harmony is by investigating alignment of the mathematics of a particular sound or harmony. The study examined preference responses to excerpts tuned in just intonation, Pythagorean intonation, and equal temperament. Musical excerpts were presented in pairs and study subjects simply picked one version from the pair that they perceived as the most consonant. Results of the study revealed an overall preference for equal temperament in contradiction to coincidence theory. Several additional areas for research are suggested to further investigate the results of this study.
Almeida, Andre; George, David; Smith, John; Wolfe, Joe
2013-09-01
Using an automated clarinet playing system, the frequency f, sound level L, and spectral characteristics are measured as functions of blowing pressure P and the force F applied by the mechanical lip at different places on the reed. The playing regime on the (P,F) plane lies below an extinction line F(P) with a negative slope of a few square centimeters and above a pressure threshold with a more negative slope. Lower values of F and P can produce squeaks. Over much of the playing regime, lines of equal frequency have negative slope. This is qualitatively consistent with passive reed behavior: Increasing F or P gradually closes the reed, reducing its equivalent acoustic compliance, which increases the frequency of the peaks of the parallel impedance of bore and reed. High P and low F produce the highest sound levels and stronger higher harmonics. At low P, sound level can be increased at constant frequency by increasing P while simultaneously decreasing F. At high P, where lines of equal f and of equal L are nearly parallel, this compensation is less effective. Applying F further from the mouthpiece tip moves the playing regime to higher F and P, as does a stiffer reed.
Broad band sound from wind turbine generators
NASA Technical Reports Server (NTRS)
Hubbard, H. H.; Shepherd, K. P.; Grosveld, F. W.
1981-01-01
Brief descriptions are given of the various types of large wind turbines and their sound characteristics. Candidate sources of broadband sound are identified and are rank ordered for a large upwind configuration wind turbine generator for which data are available. The rotor is noted to be the main source of broadband sound which arises from inflow turbulence and from the interactions of the turbulent boundary layer on the blade with its trailing edge. Sound is radiated about equally in all directions but the refraction effects of the wind produce an elongated contour pattern in the downwind direction.
Characterisation of structure-borne sound source using reception plate method.
Putra, A; Saari, N F; Bakri, H; Ramlan, R; Dan, R M
2013-01-01
A laboratory-based experiment procedure of reception plate method for structure-borne sound source characterisation is reported in this paper. The method uses the assumption that the input power from the source installed on the plate is equal to the power dissipated by the plate. In this experiment, rectangular plates having high and low mobility relative to that of the source were used as the reception plates and a small electric fan motor was acting as the structure-borne source. The data representing the source characteristics, namely, the free velocity and the source mobility, were obtained and compared with those from direct measurement. Assumptions and constraints employing this method are discussed.
Tinnitus retraining therapy for patients with tinnitus and decreased sound tolerance.
Jastreboff, Pawel J; Jastreboff, Margaret M
2003-04-01
Our experience has revealed the following: (1) TRT is applicable for all types of tinnitus, as well as for decreased sound tolerance, with significant improvement of tinnitus occurring in over 80% of the cases, and at least equal success rate for decreased sound tolerance. (2) TRT can provide cure for decreased sound tolerance. (3) TRT does not require frequent clinic visits and has no side effects; however, (4) Special training of health providers involved in this treatment is required for this treatment to be effective.
New acoustical technology of sound absorption based on reverse horn
NASA Astrophysics Data System (ADS)
Zhang, Yong Yan; Wu, Jiu Hui; Cao, Song Hua; Cao, Pei; Zhao, Zi Ting
2016-12-01
In this paper, a novel reverse horn’s sound-absorption mechanism and acoustic energy focusing mechanism for low-frequency broadband are presented. Due to the alternation of the reverse horn’s thickness, the amplitude of the acoustic pressure propagated in the structure changes, which results in growing energy focused in the edge and in the reverse horn’s tip when the characteristic length is equal to or less than a wavelength and the incident wave is compressed. There are two kinds of methods adopted to realize energy dissipation. On the one hand, sound-absorbing materials are added in incident direction in order to overcome the badness of the reverse horn’s absorption in high frequency and improve the overall high-frequency and low-frequency sound-absorption coefficients; on the other hand, adding mass and film in its tip could result in mechanical energy converting into heat energy due to the coupled vibration of mass and the film. Thus, the reverse horn with film in the tip could realize better sound absorption for low-frequency broadband. These excellent properties could have potential applications in the one-dimensional absorption wedge and for the control of acoustic wave.
An Inexpensive Group FM Amplification System for the Classroom.
ERIC Educational Resources Information Center
Worner, William A.
1988-01-01
An inexpensive FM amplification system was developed to enhance auditory learning in classrooms for the hearing impaired. Evaluation indicated that the system equalizes the sound pressure level throughout the room, with the increased sound pressure level falling in the range of 70 to 73 decibels. (Author/DB)
Torija, Antonio J; Ruiz, Diego P
2012-10-01
Road traffic has a heavy impact on the urban sound environment, constituting the main source of noise and widely dominating its spectral composition. In this context, our research investigates the use of recorded sound spectra as input data for the development of real-time short-term road traffic flow estimation models. For this, a series of models based on the use of Multilayer Perceptron Neural Networks, multiple linear regression, and the Fisher linear discriminant were implemented to estimate road traffic flow as well as to classify it according to the composition of heavy vehicles and motorcycles/mopeds. In view of the results, the use of the 50-400 Hz and 1-2.5 kHz frequency ranges as input variables in multilayer perceptron-based models successfully estimated urban road traffic flow with an average percentage of explained variance equal to 86%, while the classification of the urban road traffic flow gave an average success rate of 96.1%. Copyright © 2012 Elsevier B.V. All rights reserved.
Application of a finite-element model to low-frequency sound insulation in dwellings.
Maluski, S P; Gibbs, B M
2000-10-01
The sound transmission between adjacent rooms has been modeled using a finite-element method. Predicted sound-level difference gave good agreement with experimental data using a full-scale and a quarter-scale model. Results show that the sound insulation characteristics of a party wall at low frequencies strongly depend on the modal characteristics of the sound field of both rooms and of the partition. The effect of three edge conditions of the separating wall on the sound-level difference at low frequencies was examined: simply supported, clamped, and a combination of clamped and simply supported. It is demonstrated that a clamped partition provides greater sound-level difference at low frequencies than a simply supported. It also is confirmed that the sound-pressure level difference is lower in equal room than in unequal room configurations.
Western Ross Sea and McMurdo Sound Ice Forecasting Guide.
1975-06-01
areal ice distribution and follow the sane historical proqression of pack disintergration . This technique assumes that environmental conditions...30-day) are based on historical ice data which cxnbine averaae disintergration rates as well as averace wind and current drift. Iong-range wind...original 2 to 3 okta area and its new cnfiguration remains the same, the products of ocnoentrations and widths at the verifying time must equal the
40 CFR 205.54-2 - Sound data acquisition system.
Code of Federal Regulations, 2012 CFR
2012-07-01
... of 86 dB (rms) and the level indicated for an octave band of random noise of equal energy as the... Publication 179, Precision Sound Level Meters. (v) Magnetic tape recorders. No requirements are described in...) Calibrate tape recorders using the brand and type of magnetic tape used for actual data acquisition...
ERIC Educational Resources Information Center
La Brecque, Richard
This paper clarifies core concepts in a Kentucky judge's decision that the State General Assembly has failed to provide an efficient system of common schools. Connecting "efficiency" of educational systems to "equality of educational opportunity," the paper argues that the realization of a constitutionally sound, efficient…
Phase Shifting and the Beating of Complex Waves
ERIC Educational Resources Information Center
Keeports, David
2011-01-01
At the introductory level, the demonstration and analysis of sound beating is usually limited to the superposition of two purely sinusoidal waves with equal amplitudes and very similar frequencies. Under such conditions, an observer hears the periodic variation of the loudness of a sound with an unchanging timbre. On the other hand, when complex…
Differing effects of two synthetic phonics programmes on early reading development.
Shapiro, Laura R; Solity, Jonathan
2016-06-01
Synthetic phonics is the widely accepted approach for teaching reading in English: Children are taught to sound out the letters in a word then blend these sounds together. We compared the impact of two synthetic phonics programmes on early reading. Children received Letters and Sounds (L&S; 7 schools) which teaches multiple letter-sound mappings or Early Reading Research (ERR; 10 schools) which teaches only the most consistent mappings plus frequent words by sight. We measured phonological awareness (PA) and reading from school entry to the end of the second (all schools) or third school year (4 ERR, 3 L&S schools). Phonological awareness was significantly related to all reading measures for the whole sample. However, there was a closer relationship between PA and exception word reading for children receiving the L&S programme. The programmes were equally effective overall, but their impact on reading significantly interacted with school-entry PA: Children with poor PA at school entry achieved higher reading attainments under ERR (significant group difference on exception word reading at the end of the first year), whereas children with good PA performed equally well under either programme. The more intensive phonics programme (L&S) heightened the association between PA and exception word reading. Although the programmes were equally effective for most children, results indicate potential benefits of ERR for children with poor PA. We suggest that phonics programmes could be simplified to teach only the most consistent mappings plus frequent words by sight. © 2015 The British Psychological Society.
Andrés, Pilar; Parmentier, Fabrice B R; Escera, Carles
2006-01-01
The aim of this study was to examine the effects of aging on the involuntary capture of attention by irrelevant sounds (distraction) and the use of these sounds as warning cues (alertness) in an oddball paradigm. We compared the performance of older and younger participants on a well-characterized auditory-visual distraction task. Based on the dissociations observed in aging between attentional processes sustained by the anterior and posterior attentional networks, our prediction was that distraction by irrelevant novel sounds would be stronger in older adults than in young adults while both groups would be equally able to use sound as an alert to prepare for upcoming stimuli. The results confirmed both predictions: there was a larger distraction effect in the older participants, but the alert effect was equivalent in both groups. These results give support to the frontal hypothesis of aging [Raz, N. (2000). Aging of the brain and its impact on cognitive performance: integration of structural and functional finding. In F.I.M. Craik & T.A. Salthouse (Eds.) Handbook of aging and cognition (pp. 1-90). Mahwah, NJ: Erlbaum; West, R. (1996). An application of prefrontal cortex function theory to cognitive aging. Psychological Bulletin, 120, 272-292].
Linear multivariate evaluation models for spatial perception of soundscape.
Deng, Zhiyong; Kang, Jian; Wang, Daiwei; Liu, Aili; Kang, Joe Zhengyu
2015-11-01
Soundscape is a sound environment that emphasizes the awareness of auditory perception and social or cultural understandings. The case of spatial perception is significant to soundscape. However, previous studies on the auditory spatial perception of the soundscape environment have been limited. Based on 21 native binaural-recorded soundscape samples and a set of auditory experiments for subjective spatial perception (SSP), a study of the analysis among semantic parameters, the inter-aural-cross-correlation coefficient (IACC), A-weighted-equal sound-pressure-level (L(eq)), dynamic (D), and SSP is introduced to verify the independent effect of each parameter and to re-determine some of their possible relationships. The results show that the more noisiness the audience perceived, the worse spatial awareness they received, while the closer and more directional the sound source image variations, dynamics, and numbers of sound sources in the soundscape are, the better the spatial awareness would be. Thus, the sensations of roughness, sound intensity, transient dynamic, and the values of Leq and IACC have a suitable range for better spatial perception. A better spatial awareness seems to promote the preference slightly for the audience. Finally, setting SSPs as functions of the semantic parameters and Leq-D-IACC, two linear multivariate evaluation models of subjective spatial perception are proposed.
Broadband sound blocking in phononic crystals with rotationally symmetric inclusions.
Lee, Joong Seok; Yoo, Sungmin; Ahn, Young Kwan; Kim, Yoon Young
2015-09-01
This paper investigates the feasibility of broadband sound blocking with rotationally symmetric extensible inclusions introduced in phononic crystals. By varying the size of four equally shaped inclusions gradually, the phononic crystal experiences remarkable changes in its band-stop properties, such as shifting/widening of multiple Bragg bandgaps and evolution to resonance gaps. Necessary extensions of the inclusions to block sound effectively can be determined for given incident frequencies by evaluating power transmission characteristics. By arraying finite dissimilar unit cells, the resulting phononic crystal exhibits broadband sound blocking from combinational effects of multiple Bragg scattering and local resonances even with small-numbered cells.
Keshavarz, Behrang; Campos, Jennifer L; DeLucia, Patricia R; Oberfeld, Daniel
2017-04-01
Estimating time to contact (TTC) involves multiple sensory systems, including vision and audition. Previous findings suggested that the ratio of an object's instantaneous optical size/sound intensity to its instantaneous rate of change in optical size/sound intensity (τ) drives TTC judgments. Other evidence has shown that heuristic-based cues are used, including final optical size or final sound pressure level. Most previous studies have used decontextualized and unfamiliar stimuli (e.g., geometric shapes on a blank background). Here we evaluated TTC estimates by using a traffic scene with an approaching vehicle to evaluate the weights of visual and auditory TTC cues under more realistic conditions. Younger (18-39 years) and older (65+ years) participants made TTC estimates in three sensory conditions: visual-only, auditory-only, and audio-visual. Stimuli were presented within an immersive virtual-reality environment, and cue weights were calculated for both visual cues (e.g., visual τ, final optical size) and auditory cues (e.g., auditory τ, final sound pressure level). The results demonstrated the use of visual τ as well as heuristic cues in the visual-only condition. TTC estimates in the auditory-only condition, however, were primarily based on an auditory heuristic cue (final sound pressure level), rather than on auditory τ. In the audio-visual condition, the visual cues dominated overall, with the highest weight being assigned to visual τ by younger adults, and a more equal weighting of visual τ and heuristic cues in older adults. Overall, better characterizing the effects of combined sensory inputs, stimulus characteristics, and age on the cues used to estimate TTC will provide important insights into how these factors may affect everyday behavior.
On the Lighthill relationship and sound generation from isotropic turbulence
NASA Technical Reports Server (NTRS)
Zhou, YE; Praskovsky, Alexander; Oncley, Steven
1994-01-01
In 1952, Lighthill developed a theory for determining the sound generated by a turbulent motion of a fluid. With some statistical assumptions, Proudman applied this theory to estimate the acoustic power of isotropic turbulence. Recently, Lighthill established a simple relationship that relates the fourth-order retarded time and space covariance of his stress tensor to the corresponding second-order covariance and the turbulent flatness factor, without making statistical assumptions for a homogeneous turbulence. Lilley revisited Proudman's work and applied the Lighthill relationship to evaluate directly the radiated acoustic power from isotropic turbulence. After choosing the time separation dependence in the two-point velocity time and space covariance based on the insights gained from direct numerical simulations, Lilley concluded that the Proudman constant is determined by the turbulent flatness factor and the second-order spatial velocity covariance. In order to estimate the Proudman constant at high Reynolds numbers, we analyzed a unique data set of measurements in a large wind tunnel and atmospheric surface layer that covers a range of the Taylor microscale based on Reynolds numbers 2.0 x 10(exp 3) less than or equal to R(sub lambda) less than or equal to 12.7 x 10(exp 3). Our measurements demonstrate that the Lighthill relationship is a good approximation, providing additional support to Lilley's approach. The flatness factor is found between 2.7 - 3.3 and the second order spatial velocity covariance is obtained. Based on these experimental data, the Proudman constant is estimated to be 0.68 - 3.68.
2010-01-01
Background The present study compares the value of additional use of computer simulated heart sounds, to conventional bedside auscultation training, on the cardiac auscultation skills of 3rd year medical students at Oslo University Medical School. Methods In addition to their usual curriculum courses, groups of seven students each were randomized to receive four hours of additional auscultation training either employing a computer simulator system or adding on more conventional bedside training. Cardiac auscultation skills were afterwards tested using live patients. Each student gave a written description of the auscultation findings in four selected patients, and was rewarded from 0-10 points for each patient. Differences between the two study groups were evaluated using student's t-test. Results At the auscultation test no significant difference in mean score was found between the students who had used additional computer based sound simulation compared to additional bedside training. Conclusions Students at an early stage of their cardiology training demonstrated equal performance of cardiac auscultation whether they had received an additional short auscultation course based on computer simulated training, or had had additional bedside training. PMID:20082701
Speaker Invariance for Phonetic Information: an fMRI Investigation
Salvata, Caden; Blumstein, Sheila E.; Myers, Emily B.
2012-01-01
The current study explored how listeners map the variable acoustic input onto a common sound structure representation while being able to retain phonetic detail to distinguish among the identity of talkers. An adaptation paradigm was utilized to examine areas which showed an equal neural response (equal release from adaptation) to phonetic change when spoken by the same speaker and when spoken by two different speakers, and insensitivity (failure to show release from adaptation) when the same phonetic input was spoken by a different speaker. Neural areas which showed speaker invariance were located in the anterior portion of the middle superior temporal gyrus bilaterally. These findings provide support for the view that speaker normalization processes allow for the translation of a variable speech input to a common abstract sound structure. That this process appears to occur early in the processing stream, recruiting temporal structures, suggests that this mapping takes place prelexically, before sound structure input is mapped on to lexical representations. PMID:23264714
First results from the energetic particle instrument on the OEDIPUS-C sounding rocket
NASA Astrophysics Data System (ADS)
Gough, M. P.; Hardy, D. A.; James, H. G.
The Canadian / US OEDIPUS-C rocket was flown from the Poker Flat Rocket Range November 6th 1995 as a mother-son sounding rocket. It was designed to study auroral ionospheric plasma physics using active wave sounding and prove tether technology. The payload separated into two sections reaching a separation of 1200m along the Earth's magnetic field. One section included a frequency stepped HF transmitter and the other included a synchronised HF receiver. Both sections included Energetic Particle Instruments, EPI, stepped in energy synchronously with the transmitter steps. On-board EPI particle processing in both payloads provided direct measurements of electron heating, wave-particle interactions via particle correlators, and a high resolution measurement of wave induced particle heating via transmitter synchronised fast sampling. Strong electron heating was observed at times when the HF transmitter frequency was equal to a harmonic of the electron gyrofrequency, f_ce, or equal to the upper hybrid frequency, f_uh.
Thaler, Lore; Castillo-Serrano, Josefina
2016-01-01
Echolocation is the ability to use reflected sound to obtain information about the spatial environment. Echolocation is an active process that requires both the production of the emission as well as the sensory processing of the resultant sound. Appreciating the general usefulness of echo-acoustic cues for people, in particular those with vision impairments, various devices have been built that exploit the principle of echolocation to obtain and provide information about the environment. It is common to all these devices that they do not require the person to make a sound. Instead, the device produces the emission autonomously and feeds a resultant sound back to the user. Here we tested if echolocation performance in a simple object detection task was affected by the use of a head-mounted loudspeaker as compared to active clicking. We found that 27 sighted participants new to echolocation did generally better when they used a loudspeaker as compared to mouth-clicks, and that two blind participants with experience in echolocation did equally well with mouth clicks and the speaker. Importantly, performance of sighted participants' was not statistically different from performance of blind experts when they used the speaker. Based on acoustic click data collected from a subset of our participants, those participants whose mouth clicks were more similar to the speaker clicks, and thus had higher peak frequencies and sound intensity, did better. We conclude that our results are encouraging for the consideration and development of assistive devices that exploit the principle of echolocation.
Thaler, Lore; Castillo-Serrano, Josefina
2016-01-01
Echolocation is the ability to use reflected sound to obtain information about the spatial environment. Echolocation is an active process that requires both the production of the emission as well as the sensory processing of the resultant sound. Appreciating the general usefulness of echo-acoustic cues for people, in particular those with vision impairments, various devices have been built that exploit the principle of echolocation to obtain and provide information about the environment. It is common to all these devices that they do not require the person to make a sound. Instead, the device produces the emission autonomously and feeds a resultant sound back to the user. Here we tested if echolocation performance in a simple object detection task was affected by the use of a head-mounted loudspeaker as compared to active clicking. We found that 27 sighted participants new to echolocation did generally better when they used a loudspeaker as compared to mouth-clicks, and that two blind participants with experience in echolocation did equally well with mouth clicks and the speaker. Importantly, performance of sighted participants’ was not statistically different from performance of blind experts when they used the speaker. Based on acoustic click data collected from a subset of our participants, those participants whose mouth clicks were more similar to the speaker clicks, and thus had higher peak frequencies and sound intensity, did better. We conclude that our results are encouraging for the consideration and development of assistive devices that exploit the principle of echolocation. PMID:27135407
Hearing and Underwater Noise Exposure
1985-08-27
Ordnance Laboratory, White Oak, Silver Springs, Md. Corso, J. F. (1963) Bone-conduction thresholds for sonic and ultrasonic frequencies. J. Acoust...Soc. Am. 35, 1738-1743. Corso, J. F., & Levine, M. (1965a) Sonic and Ultrasonic equal loudness contours. J. Exp. Psychol. 70, 412-416. Corso, J...KEY WORDS (Continue on reverse aide II neceaaary and identity by block number) Hearing, Deafness, Sonar, Noise, Sound, Ultrasound, Infra -sound
Apparatus and method for processing Korotkov sounds. [for blood pressure measurement
NASA Technical Reports Server (NTRS)
Golden, D. P., Jr.; Hoffler, G. W.; Wolthuis, R. A. (Inventor)
1974-01-01
A Korotkov sound processor, used in a noninvasive automatic blood measuring system where the brachial artery is occluded by an inflatable cuff, is disclosed. The Korotkoff sound associated with the systolic event is determined when the ratio of the absolute value of a voltage signal, representing Korotkov sounds in the range of 18 to 26 Hz to a maximum absolute peak value of the unfiltered signals, first equals or exceeds a value of 0.45. Korotkov sound associated with the diastolic event is determined when a ratio of the voltage signal of the Korotkov sounds in the range of 40 to 60 Hz to the absolute peak value of such signals within a single measurement cycle first falls below a value of 0.17. The processor signals the occurrence of the systolic and diastolic events and these signals can be used to control a recorder to record pressure values for these events.
Power-output regularization in global sound equalization.
Stefanakis, Nick; Sarris, John; Cambourakis, George; Jacobsen, Finn
2008-01-01
The purpose of equalization in room acoustics is to compensate for the undesired modification that an enclosure introduces to signals such as audio or speech. In this work, equalization in a large part of the volume of a room is addressed. The multiple point method is employed with an acoustic power-output penalty term instead of the traditional quadratic source effort penalty term. Simulation results demonstrate that this technique gives a smoother decline of the reproduction performance away from the control points.
The measurement of Eustachian tube function in a hyperbaric chamber using an ear canal microphone.
Fischer, Hans-Georg; Koch, Andreas; Kähler, Wataru; Pohl, Michael; Pau, Hans-Wilhelm; Zehlicke, Thorsten
2016-03-01
The purpose of this study was to further the understanding of the opening of the Eustachian tube in relation to changes in barometric pressure. An ear canal microphone was used to measure the specific sounds related to tube opening and possible eardrum movements. Five subjects with normal tube function were examined in a hyperbaric chamber (up to 304 kPa). All active and passive equalization events were recorded and correlated with the subjectively perceived pressure regulation in the measured ear. The signals recorded were clear and reproducible. The acoustic analysis distinguished between the different kinds of equalization. Subjective impressions were confirmed by the recorded frequency of acoustic phenomena (clicks). During compression, the sequence of active equalization manoeuvres was in a more regular and steady pattern than during decompression, when the click sounds varied. The study established a simple technical method for analyzing the function of the Eustachian tube and provided new information about barometric pressure regulation of the middle ear.
NASA Astrophysics Data System (ADS)
Chabassier, Juliette; Duruflé, Marc
2014-12-01
A nonlinear model for a vibrating Timoshenko beam in non-forced unknown rotation is derived from the virtual work principle applied to a system of beam with mass at the end. The system represents a piano hammer shank coupled to a hammer head. An energy-based numerical scheme is then provided, obtained by non-classical approaches. A major difficulty for time discretization comes from the nonlinear behavior of the kinetic energy of the system. This new numerical scheme is then coupled to a global energy-preserving numerical solution for the whole piano. The obtained numerical simulations show that the pianistic touch clearly influences the spectrum of the piano sound of equally loud isolated notes. These differences do not come from a possible shock excitation on the structure, or from a changing impact point, or a “longitudinal rubbing motion” on the string, since neither of these features is modeled in our study.
Human sensorimotor tracking of continuous subliminal deviations from isochrony.
Madison, Guy; Merker, Björn
2004-11-03
We show that people continuously react to time perturbations in the range 3-96 ms in otherwise isochronous sound sequences. Musically trained and untrained participants were asked to synchronize with a sequence of sounds, and these two groups performed almost equally below the threshold for conscious detection of the perturbations. Above this threshold the motor reactions accounted for a larger proportion of the stimulus deviations in musically trained participants.
NASA Technical Reports Server (NTRS)
Powell, C. A.; Mccurdy, D. A.
1978-01-01
Fifty-two airplane noise recordings, made at several locations around Dulles International Airport, were analyzed to compare the low-frequency noise levels of the Concorde supersonic transport with those of other commercial jet airplanes. Comparisons of the relative low-frequency noise levels which were produced at close and distant locations for departures and arrivals were made for three noise measures: the sound pressure level in the 1/3 octave band centered at 20 Hz, the total sound pressure level in the 1/3 octave bands with center frequencies less than or equal to 125 Hz, and the total sound pressure level in the 1/3 octave bands with center frequencies less than or equal to 500 Hz. Although the absolute noise levels for Concorde were found, in general, to be higher than those for the other airplane types, the level of low-frequency noise of the Concorde relative to the perceived noise level (PNL), effective perceived noise level (EPNL), and overall sound pressure level (OASPL) was within the range established by the other airplane types, except for the arrival operations of four-engine, narrow-body airplanes. The measure OASPL was found to be a significantly better predictor of low-frequency noise level than PNL or EPNL.
The effect of spatial distribution on the annoyance caused by simultaneous sounds
NASA Astrophysics Data System (ADS)
Vos, Joos; Bronkhorst, Adelbert W.; Fedtke, Thomas
2004-05-01
A considerable part of the population is exposed to simultaneous and/or successive environmental sounds from different sources. In many cases, these sources are different with respect to their locations also. In a laboratory study, it was investigated whether the annoyance caused by the multiple sounds is affected by the spatial distribution of the sources. There were four independent variables: (1) sound category (stationary or moving), (2) sound type (stationary: lawn-mower, leaf-blower, and chain saw; moving: road traffic, railway, and motorbike), (3) spatial location (left, right, and combinations), and (4) A-weighted sound exposure level (ASEL of single sources equal to 50, 60, or 70 dB). In addition to the individual sounds in isolation, various combinations of two or three different sources within each sound category and sound level were presented for rating. The annoyance was mainly determined by sound level and sound source type. In most cases there were neither significant main effects of spatial distribution nor significant interaction effects between spatial distribution and the other variables. It was concluded that for rating the spatially distrib- uted sounds investigated, the noise dose can simply be determined by a summation of the levels for the left and right channels. [Work supported by CEU.
Not all carp are created equal: Impacts of broadband sound on common carp swimming behavior
Murchy, Kelsie; Vetter, Brooke J.; Brey, Marybeth; Amberg, Jon J.; Gaikowski, Mark; Mensinger, Allen F.
2016-01-01
Bighead carp (Hypophthalmichthys nobilis), silver carp (H. molitrix) (hereafter: bigheaded carps), and common carp (Cyprinus carpio) are invasive fish causing negative impacts throughout their North American range. To control their movements, non-physical barriers are being developed. Broadband sound (0.06 to 10 kHz) has shown potential as an acoustic deterrent for bigheaded carps, but the response of common carp to broadband sound has not been evaluated. Since common carp are ostariophysians, possessing Weberian ossicles similar to bigheaded carps, it is possible that sound can be used as an acoustical deterrent for all three species. Behavioral responses to a broadband sound were evaluated for common carp in an outdoor concrete pond. Common carp responded a median of 3.0 (1st Q: 1.0, 3rd Q: 6.0) consecutive times to the broadband sound which was lower than silver carp and bighead carp to the same stimulus. The current study shows that common carp demonstrate an inconsistent negative phonotaxis response to a broadband sound, and seem to habituate to the sound quickly.
Chen, Yi-Chuan; Huang, Pi-Chun; Woods, Andy; Spence, Charles
2016-05-27
It has been suggested that the Bouba/Kiki effect, in which meaningless speech sounds are systematically mapped onto rounded or angular shapes, reflects a universal crossmodal correspondence between audition and vision. Here, radial frequency (RF) patterns were adapted in order to compare the Bouba/Kiki effect in Eastern and Western participants demonstrating different perceptual styles. Three attributes of the RF patterns were manipulated: The frequency, amplitude, and spikiness of the sinusoidal modulations along the circumference of a circle. By testing participants in the US and Taiwan, both cultural commonalities and differences in sound-shape correspondence were revealed. RF patterns were more likely to be matched with "Kiki" than with "Bouba" when the frequency, amplitude, and spikiness increased. The responses from both groups of participants had a similar weighting on frequency; nevertheless, the North Americans had a higher weighting on amplitude, but a lower weighting on spikiness, than their Taiwanese counterparts. These novel results regarding cultural differences suggest that the Bouba/Kiki effect is partly tuned by differing perceptual experience. In addition, using the RF patterns in the Bouba/Kiki effect provides a "mid-level" linkage between visual and auditory processing, and a future understanding of sound-shape correspondences based on the mechanism of visual pattern processing.
On the upper bound in the Bohm sheath criterion
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kotelnikov, I. A., E-mail: I.A.Kotelnikov@inp.nsk.su; Skovorodin, D. I., E-mail: D.I.Skovorodin@inp.nsk.su
2016-02-15
The question is discussed about the existence of an upper bound in the Bohm sheath criterion, according to which the Debye sheath at the interface between plasma and a negatively charged electrode is stable only if the ion flow velocity in plasma exceeds the ion sound velocity. It is stated that, with an exception of some artificial ionization models, the Bohm sheath criterion is satisfied as an equality at the lower bound and the ion flow velocity is equal to the speed of sound. In the one-dimensional theory, a supersonic flow appears in an unrealistic model of a localized ionmore » source the size of which is less than the Debye length; however, supersonic flows seem to be possible in the two- and three-dimensional cases. In the available numerical codes used to simulate charged particle sources with a plasma emitter, the presence of the upper bound in the Bohm sheath criterion is not supposed; however, the correspondence with experimental data is usually achieved if the ion flow velocity in plasma is close to the ion sound velocity.« less
Wireless and acoustic hearing with bone-anchored hearing devices.
Bosman, Arjan J; Mylanus, Emmanuel A M; Hol, Myrthe K S; Snik, Ad F M
2015-07-01
The efficacy of wireless connectivity in bone-anchored hearing was studied by comparing the wireless and acoustic performance of the Ponto Plus sound processor from Oticon Medical relative to the acoustic performance of its predecessor, the Ponto Pro. Nineteen subjects with more than two years' experience with a bone-anchored hearing device were included. Thirteen subjects were fitted unilaterally and six bilaterally. Subjects served as their own control. First, subjects were tested with the Ponto Pro processor. After a four-week acclimatization period performance the Ponto Plus processor was measured. In the laboratory wireless and acoustic input levels were made equal. In daily life equal settings of wireless and acoustic input were used when watching TV, however when using the telephone the acoustic input was reduced by 9 dB relative to the wireless input. Speech scores for microphone with Ponto Pro and for both input modes of the Ponto Plus processor were essentially equal when equal input levels of wireless and microphone inputs were used. Only the TV-condition showed a statistically significant (p <5%) lower speech reception threshold for wireless relative to microphone input. In real life, evaluation of speech quality, speech intelligibility in quiet and noise, and annoyance by ambient noise, when using landline phone, mobile telephone, and watching TV showed a clear preference (p <1%) for the Ponto Plus system with streamer over the microphone input. Due to the small number of respondents with landline phone (N = 7) the result for noise annoyance was only significant at the 5% level. Equal input levels for acoustic and wireless inputs results in equal speech scores, showing a (near) equivalence for acoustic and wireless sound transmission with Ponto Pro and Ponto Plus. The default 9-dB difference between microphone and wireless input when using the telephone results in a substantial wireless benefit when using the telephone. The preference of wirelessly transmitted audio when watching TV can be attributed to the relatively poor sound quality of backward facing loudspeakers in flat screen TVs. The ratio of wireless and acoustic input can be easily set to the user's preference with the streamer's volume control.
Beranek, Leo
2011-05-01
The parameter, "Strength of Sound G" is closely related to loudness. Its magnitude is dependent, inversely, on the total sound absorption in a room. By comparison, the reverberation time (RT) is both inversely related to the total sound absorption in a hall and directly related to its cubic volume. Hence, G and RT in combination are vital in planning the acoustics of a concert hall. A newly proposed "Bass Index" is related to the loudness of the bass sound and equals the value of G at 125 Hz in decibels minus its value at mid-frequencies. Listener envelopment (LEV) is shown for most halls to be directly related to the mid-frequency value of G. The broadening of sound, i.e., apparent source width (ASW) is given by degree of source broadening (DSB) which is determined from the combined effect of early lateral reflections as measured by binaural quality index (BQI) and strength G. The optimum values and limits of these parameters are discussed.
NASA Astrophysics Data System (ADS)
Liang, Ruiyu; Xi, Ji; Bao, Yongqiang
2017-07-01
To improve the performance of gain compensation based on three-segment sound pressure level (SPL) in hearing aids, an improved multichannel loudness compensation method based on eight-segment SPL was proposed. Firstly, the uniform cosine modulated filter bank was designed. Then, the adjacent channels which have low or gradual slopes were adaptively merged to obtain the corresponding non-uniform cosine modulated filter according to the audiogram of hearing impaired persons. Secondly, the input speech was decomposed into sub-band signals and the SPL of every sub-band signal was computed. Meanwhile, the audible SPL range from 0 dB SPL to 120 dB SPL was equally divided into eight segments. Based on these segments, a different prescription formula was designed to compute more detailed gain to compensate according to the audiogram and the computed SPL. Finally, the enhanced signal was synthesized. Objective experiments showed the decomposed signals after cosine modulated filter bank have little distortion. Objective experiments showed that the hearing aids speech perception index (HASPI) and hearing aids speech quality index (HASQI) increased 0.083 and 0.082 on average, respectively. Subjective experiments showed the proposed algorithm can effectively improve the speech recognition of six hearing impaired persons.
On the Possible Detection of Lightning Storms by Elephants
Kelley, Michael C.; Garstang, Michael
2013-01-01
Simple Summary We use data similar to that taken by the International Monitoring System for the detection of nuclear explosions, to determine whether elephants might be capable of detecting and locating the source of sounds generated by thunderstorms. Knowledge that elephants might be capable of responding to such storms, particularly at the end of the dry season when migrations are initiated, is of considerable interest to management and conservation. Abstract Theoretical calculations suggest that sounds produced by thunderstorms and detected by a system similar to the International Monitoring System (IMS) for the detection of nuclear explosions at distances ≥100 km, are at sound pressure levels equal to or greater than 6 × 10−3 Pa. Such sound pressure levels are well within the range of elephant hearing. Frequencies carrying these sounds might allow for interaural time delays such that adult elephants could not only hear but could also locate the source of these sounds. Determining whether it is possible for elephants to hear and locate thunderstorms contributes to the question of whether elephant movements are triggered or influenced by these abiotic sounds. PMID:26487406
Liquefaction potential index: Field assessment
Toprak, S.; Holzer, T.L.
2003-01-01
Cone penetration test (CPT) soundings at historic liquefaction sites in California were used to evaluate the predictive capability of the liquefaction potential index (LPI), which was defined by Iwasaki et al. in 1978. LPI combines depth, thickness, and factor of safety of liquefiable material inferred from a CPT sounding into a single parameter. LPI data from the Monterey Bay region indicate that the probability of surface manifestations of liquefaction is 58 and 93%, respectively, when LPI equals or exceeds 5 and 15. LPI values also generally correlate with surface effects of liquefaction: Decreasing from a median of 12 for soundings in lateral spreads to 0 for soundings where no surface effects were reported. The index is particularly promising for probabilistic liquefaction hazard mapping where it may be a useful parameter for characterizing the liquefaction potential of geologic units.
Bonfiglio, Paolo; Pompoli, Francesco; Lionti, Riccardo
2016-04-01
The transfer matrix method is a well-established prediction tool for the simulation of sound transmission loss and the sound absorption coefficient of flat multilayer systems. Much research has been dedicated to enhancing the accuracy of the method by introducing a finite size effect of the structure to be simulated. The aim of this paper is to present a reduced-order integral formulation to predict radiation efficiency and radiation impedance for a panel with equal lateral dimensions. The results are presented and discussed for different materials in terms of radiation efficiency, sound transmission loss, and the sound absorption coefficient. Finally, the application of the proposed methodology for rectangular multilayer systems is also investigated and validated against experimental data.
Sound Radiation from a Supersonic Jet Passing Through a Partially Open Exhaust Duct
NASA Technical Reports Server (NTRS)
Kandula, Max
2011-01-01
The radiation of sound from a perfectly expanded Mach 2.5 cold supersonic jet of 25.4 mm exit diameter flowing through a partially open rigid-walled duct with an upstream i-deflector has been studied experimentally. In the experiments, the nozzle is mounted vertically, with the nozzle exit plane at a height of 73 jet diameters above ground level. Relative to the nozzle exit plane (NEP), the location of the duct inlet is varied at 10, 5, and -1 jet diameters. Far-field sound pressure levels were obtained at 54 jet diameters above ground with the aid of acoustic sensors equally spaced around a circular arc of radius equal to 80 jet diameters from the jet axis. Data on the jet acoustic field for the partially open duct were obtained and compared with those with a free jet and with a closed duct. The results suggest that for the partially open duct the overall sound pressure level (OASPL) decreases as the distance between the NEP and the duct inlet plane decreases, while the opposite trend is observed for the closed duct. It is also concluded that the observed peak frequency in the partially open duct increases above the free jet value as the angle from the duct axis is increased, and as the duct inlet plane becomes closer to the NEP.
NASA Technical Reports Server (NTRS)
Rasbury, Jack; Bascal, Kira; Ownby, Matt; McCulley, Phyllis; Paul, Bonnie
2004-01-01
A traditional stethoscope is currently flown on the International Space Station (ISS). The background noise on the ISS is much higher than a normal exam room, and the literature shows that traditional stethoscopes are unable to function effectively in high noise environments. Digital stethoscopes provide amplification which improves the audibility in a quiet environment. This study is designed to determine if digital stethoscopes offer any advantage over traditional stethoscopes in being able to identify normal and abnormal sounds in the ISS noise environment. Methods: An ISS noise simulation facility was created to reproduce ISS noise profiles by modifying pink noise with a software-based graphic equalizer. The files were played in a continuous loop on a computer, amplified through a high-end stereo system and adjusted using a sound level meter. Nine caregiver analogues were given the same auscultation lesson received by astronauts. They began testing by becoming familiar with normal and abnormal sounds on a Student Auscultation Manikin . They then used two digital stethoscopes and a traditional stethoscope identical to the one flown on the ISS to auscultate the manikin sounds in the noise facility. They identified the sounds on a questionnaire and picked which of the three stethoscopes they preferred. Results: Evaluators displayed equivalent accuracy in sound identification when using either the 3M model 4000 digital stethoscope or traditional stethoscope. However, the 3M was preferred 2 to 1 by the evaluators, primarily because of additional amplification of the sounds. Discussion: Although our results show that the current ISS stethoscope and the "best-of-breed" digital stethoscope provide essentially the same auscultation utility, the latter has the advantage of recording and transmitting sounds to a remote physician. Since the astronaut caregivers are non-physiCians, this capability may be worth the additional expense and effort needed to certify the digital stethoscope for flight.
Spatial traffic noise pollution assessment - A case study.
Monazzam, Mohammad Reza; Karimi, Elham; Abbaspour, Majid; Nassiri, Parvin; Taghavi, Lobat
2015-01-01
Spatial assessment of traffic noise pollution intensity will provide urban planners with approximate estimation of citizens exposure to impermissible sound levels. They could identify critical noise pollution areas wherein noise barriers should be embedded. The present study aims at using the Geographic Information System (GIS) to assess spatial changes in traffic noise pollution in Tehran, the capital of Iran, and the largest city in the Middle East. For this purpose, while measuring equivalent sound levels at different time periods of a day and different days of a week in District 14 of Tehran, wherein there are highways and busy streets, the geographic coordination of the measurement points was recorded at the stations. The obtained results indicated that the equivalent sound level did not show a statistically significant difference between weekdays, and morning, afternoon and evening hours as well as time intervals of 10 min, 15 min and 30 min. Then, 91 stations were selected in the target area and equivalent sound level was measured for each station on 3 occasions of the morning (7:00-9:00 a.m.), afternoon (12.00-3:00 p.m.) and evening (5:00-8:00 p.m.) on Saturdays to Wednesdays. As the results suggest, the maximum equivalent sound level (Leq) was reported from Basij Highway, which is a very important connecting thoroughfare in the district, and was equal to 84.2 dB(A), while the minimum equivalent sound level (Leq), measured in the Fajr Hospital, was equal to 59.9 dB(A). The average equivalent sound level was higher than the national standard limit at all stations. The use of sound walls in Highways Basij and Mahallati as well as widening the Streets 17th Shahrivar, Pirouzi and Khavaran, benchmarked on a map, were recommended as the most effective mitigation measures. Additionally, the research findings confirm the outstanding applicability of the Geographic Information System in handling noise pollution data towards depicting noise pollution intensity caused by traffic. This work is available in Open Access model and licensed under a CC BY-NC 3.0 PL license.
Exact solution of the Lifshitz equations governing the growth of fluctuations in cosmology
NASA Technical Reports Server (NTRS)
Adams, P. J.; Canuto, V.
1975-01-01
The exact solution of the Lifshitz equations governing the cosmological evolution of an initial fluctuation is presented. Lifshitz results valid for squares of the sound velocity equal to zero and 1/3 are extended in closed form to any equation of state where the pressure equals the total energy density times the square of the sound velocity. The solutions embody all the results found previously for special cases of the square of the sound velocity. It is found that the growth of any initial fluctuation is only an exponential function of time with an exponent of not more than 4/3 and is insufficient to produce galaxies unless the initial fluctuation is very large. A possible way to produce very large initial fluctuations by modifying the equation of state by including gravitational interactions is also examined. It is found that a phase transition can occur at baryonic density of 1 nucleon per cubic Planck length or equivalently, at a time of about 10 to the -43rd power sec. At those early times, the masses allowed by causality requirements are too small to be of interest in galaxy formation.
Vibration analysis and sound field characteristics of a tubular ultrasonic radiator.
Liang, Zhaofeng; Zhou, Guangping; Zhang, Yihui; Li, Zhengzhong; Lin, Shuyu
2006-12-01
A sort of tubular ultrasonic radiator used in ultrasonic liquid processing is studied. The frequency equation of the tubular radiator is derived, and its radiated sound field in cylindrical reactor is calculated using finite element method and recorded by means of aluminum foil erosion. The results indicate that sound field of tubular ultrasonic radiator in cylindrical reactor appears standing waves along both its radial direction and axial direction, and amplitudes of standing waves decrease gradually along its radial direction, and the numbers of standing waves along its axial direction are equal to the axial wave numbers of tubular radiator. The experimental results are in good agreement with calculated results.
Kocsis, Zsuzsanna; Winkler, István; Bendixen, Alexandra; Alain, Claude
2016-09-01
The auditory environment typically comprises several simultaneously active sound sources. In contrast to the perceptual segregation of two concurrent sounds, the perception of three simultaneous sound objects has not yet been studied systematically. We conducted two experiments in which participants were presented with complex sounds containing sound segregation cues (mistuning, onset asynchrony, differences in frequency or amplitude modulation or in sound location), which were set up to promote the perceptual organization of the tonal elements into one, two, or three concurrent sounds. In Experiment 1, listeners indicated whether they heard one, two, or three concurrent sounds. In Experiment 2, participants watched a silent subtitled movie while EEG was recorded to extract the object-related negativity (ORN) component of the event-related potential. Listeners predominantly reported hearing two sounds when the segregation promoting manipulations were applied to the same tonal element. When two different tonal elements received manipulations promoting them to be heard as separate auditory objects, participants reported hearing two and three concurrent sounds objects with equal probability. The ORN was elicited in most conditions; sounds that included the amplitude- or the frequency-modulation cue generated the smallest ORN amplitudes. Manipulating two different tonal elements yielded numerically and often significantly smaller ORNs than the sum of the ORNs elicited when the same cues were applied on a single tonal element. These results suggest that ORN reflects the presence of multiple concurrent sounds, but not their number. The ORN results are compatible with the horse-race principle of combining different cues of concurrent sound segregation. Copyright © 2016 Elsevier B.V. All rights reserved.
A Survey on the Feasibility of Sound Classification on Wireless Sensor Nodes
Salomons, Etto L.; Havinga, Paul J. M.
2015-01-01
Wireless sensor networks are suitable to gain context awareness for indoor environments. As sound waves form a rich source of context information, equipping the nodes with microphones can be of great benefit. The algorithms to extract features from sound waves are often highly computationally intensive. This can be problematic as wireless nodes are usually restricted in resources. In order to be able to make a proper decision about which features to use, we survey how sound is used in the literature for global sound classification, age and gender classification, emotion recognition, person verification and identification and indoor and outdoor environmental sound classification. The results of the surveyed algorithms are compared with respect to accuracy and computational load. The accuracies are taken from the surveyed papers; the computational loads are determined by benchmarking the algorithms on an actual sensor node. We conclude that for indoor context awareness, the low-cost algorithms for feature extraction perform equally well as the more computationally-intensive variants. As the feature extraction still requires a large amount of processing time, we present four possible strategies to deal with this problem. PMID:25822142
Optical measurement of sound using time-varying laser speckle patterns
NASA Astrophysics Data System (ADS)
Leung, Terence S.; Jiang, Shihong; Hebden, Jeremy
2011-02-01
In this work, we introduce an optical technique to measure sound. The technique involves pointing a coherent pulsed laser beam on the surface of the measurement site and capturing the time-varying speckle patterns using a CCD camera. Sound manifests itself as vibrations on the surface which induce a periodic translation of the speckle pattern over time. Using a parallel speckle detection scheme, the dynamics of the time-varying speckle patterns can be captured and processed to produce spectral information of the sound. One potential clinical application is to measure pathological sounds from the brain as a screening test. We performed experiments to demonstrate the principle of the detection scheme using head phantoms. The results show that the detection scheme can measure the spectra of single frequency sounds between 100 and 2000 Hz. The detection scheme worked equally well in both a flat geometry and an anatomical head geometry. However, the current detection scheme is too slow for use in living biological tissues which has a decorrelation time of a few milliseconds. Further improvements have been suggested.
The comparative evidence for urban species sorting by anthropogenic noise
Hu, Yang
2018-01-01
Anthropogenic noise is more intense at lower sound frequencies, which could decrease urban tolerance of animals with low-frequency vocalizations. Four large comparative studies tested whether anthropogenic noise filters bird species according to the sound frequencies they use and produced discrepant results. We reanalysed data from these studies to explain their different results. Urban tolerance of bird species (defined here as often occurring and breeding in cities) is very weakly related to urban preference or relative abundance (defined based on changes in population density from urban to nearby rural environments). Data on urban preference/abundance are potentially accurate for individual cities but differ among cities for the same species, whereas existing data on urban tolerance are coarser but provide a more global synthesis. Cross-species comparisons find a positive association between the sound frequency of song and urban tolerance, but not urban preference/abundance. We found that showing an association between song frequency and urban tolerance requires controlling for additional species traits that influence urban living. On the contrary, controlling for other species traits is not required to show a positive association between song frequency and use of noisy relative to quiet areas within the same type of environment. Together, comparative evidence indicates that masking by urban noise is part of a larger set of factors influencing urban living: all else being equal, species with high-frequency sounds are more likely to tolerate cities than species with low-frequency sounds, but they are not more likely to prefer, or to be more abundant in, urban than non-urban habitats. PMID:29515901
The comparative evidence for urban species sorting by anthropogenic noise.
Cardoso, Gonçalo C; Hu, Yang; Francis, Clinton D
2018-02-01
Anthropogenic noise is more intense at lower sound frequencies, which could decrease urban tolerance of animals with low-frequency vocalizations. Four large comparative studies tested whether anthropogenic noise filters bird species according to the sound frequencies they use and produced discrepant results. We reanalysed data from these studies to explain their different results. Urban tolerance of bird species (defined here as often occurring and breeding in cities) is very weakly related to urban preference or relative abundance (defined based on changes in population density from urban to nearby rural environments). Data on urban preference/abundance are potentially accurate for individual cities but differ among cities for the same species, whereas existing data on urban tolerance are coarser but provide a more global synthesis. Cross-species comparisons find a positive association between the sound frequency of song and urban tolerance, but not urban preference/abundance. We found that showing an association between song frequency and urban tolerance requires controlling for additional species traits that influence urban living. On the contrary, controlling for other species traits is not required to show a positive association between song frequency and use of noisy relative to quiet areas within the same type of environment. Together, comparative evidence indicates that masking by urban noise is part of a larger set of factors influencing urban living: all else being equal, species with high-frequency sounds are more likely to tolerate cities than species with low-frequency sounds, but they are not more likely to prefer, or to be more abundant in, urban than non-urban habitats.
What is a melody? On the relationship between pitch and brightness of timbre.
Cousineau, Marion; Carcagno, Samuele; Demany, Laurent; Pressnitzer, Daniel
2013-01-01
Previous studies showed that the perceptual processing of sound sequences is more efficient when the sounds vary in pitch than when they vary in loudness. We show here that sequences of sounds varying in brightness of timbre are processed with the same efficiency as pitch sequences. The sounds used consisted of two simultaneous pure tones one octave apart, and the listeners' task was to make same/different judgments on pairs of sequences varying in length (one, two, or four sounds). In one condition, brightness of timbre was varied within the sequences by changing the relative level of the two pure tones. In other conditions, pitch was varied by changing fundamental frequency, or loudness was varied by changing the overall level. In all conditions, only two possible sounds could be used in a given sequence, and these two sounds were equally discriminable. When sequence length increased from one to four, discrimination performance decreased substantially for loudness sequences, but to a smaller extent for brightness sequences and pitch sequences. In the latter two conditions, sequence length had a similar effect on performance. These results suggest that the processes dedicated to pitch and brightness analysis, when probed with a sequence-discrimination task, share unexpected similarities.
Temporal processing deficit leads to impaired multisensory binding in schizophrenia.
Zvyagintsev, Mikhail; Parisi, Carmen; Mathiak, Klaus
2017-09-01
Schizophrenia has been characterised by neurodevelopmental dysconnectivity resulting in cognitive and perceptual dysmetria. Hence patients with schizophrenia may be impaired to detect the temporal relationship between stimuli in different sensory modalities. However, only a few studies described deficit in perception of temporally asynchronous multisensory stimuli in schizophrenia. We examined the perceptual bias and the processing time of synchronous and delayed sounds in the streaming-bouncing illusion in 16 patients with schizophrenia and a matched control group of 18 participants. Equal for patients and controls, the synchronous sound biased the percept of two moving squares towards bouncing as opposed to the more frequent streaming percept in the condition without sound. In healthy controls, a delay of the sound presentation significantly reduced the bias and led to prolonged processing time whereas patients with schizophrenia did not differentiate between this condition and the condition with synchronous sound. Schizophrenia leads to a prolonged window of simultaneity for audiovisual stimuli. Therefore, temporal processing deficit in schizophrenia can lead to hyperintegration of temporally unmatched multisensory stimuli.
Stanković, B; Milenković, S
2007-01-01
To compare continuous full-time (24 hours per day) occlusion of the sound eye with full-time occlusion (24 hours per day) of the sound eye 1 day more than the years of age periodically alternating with occlusion of the amblyopic eye for 1 day, as treatments for profound strabismic amblyopia in children older than 5 years. A total of 53 patients with visual acuity (VA) of 0.4 and less in the amblyopic eye (tested by crowded Landolt Cs) after previously being provided optimal optical correction were randomly assigned to receive either of the two patching regimens. VA and pattern reversal visual evoked potentials (PVEP) were recorded prospectively at 1-month intervals. Improvement in VA and the reduction in crowding difficulties (CD) were the main outcome measures of the treatment efficiency. Both treatment modalities were equally effective. Of the 51 subjects who completed the study, 21 (41.2%) were cured whereas 32 (62.7%) attained satisfactory improvement. Recovery of VA was related to age, with cure being obtained in 23.5% (4/17) and satisfactory improvement in 52.9% (9/17) of patients older than 9 years. Larger gain in VA influenced the stability of the vision over time. It can be concluded that in clinically monitored parameters both treatment modalities were equally effective without any statistical or clinical significance in the observed groups of patients. However, events like the "trade-off" effect, occurrence of occlusion amblyopia, or prolongation of PVEP latency of the sound eye indicate that full-time continuous occlusion possibly presents a more effective form of treatment.
NASA Astrophysics Data System (ADS)
Zamuraev, V. P.; Kalinina, A. P.
2017-10-01
Forced high-frequency vibrations of the airfoil surface part with the amplitude almost equal to the sound velocity can change significantly the lift force of the symmetric profile streamlined at zero angle of attack. The oscillation consists of two harmonics. The ratio of harmonics frequencies values is equal to 2. The present work shows that the aerodynamic properties depend significantly on the specific energy contribution of each frequency.
Sound levels and their effects on children in a German primary school.
Eysel-Gosepath, Katrin; Daut, Tobias; Pinger, Andreas; Lehmacher, Walter; Erren, Thomas
2012-12-01
Considerable sound levels are produced in primary schools by voices of children and resonance effects. As a consequence, hearing loss and mental impairment may occur. In a Cologne primary school, sound levels were measured in three different classrooms, each with 24 children, 8-10 years old, and one teacher. Sound dosimeters were positioned in the room and near the teacher's ear. Additional measurements were done in one classroom fully equipped with sound-absorbing materials. A questionnaire containing 12 questions about noise at school was distributed to 100 children, 8-10 years old. Measurements were repeated after children had been taught about noise damage and while "noise lights" were used. Mean sound levels of 5-h per day measuring period were 78 dB (A) near the teacher's ear and 70 dB (A) in the room. The average of all measured maximal sound levels for 1 s was 105 dB (A) for teachers, and 100 dB (A) for rooms. In the soundproofed classroom, Leq was 66 dB (A). The questionnaire revealed certain judgment of the children concerning situations with high sound levels and their ability to develop ideas for noise reduction. However, no clear sound level reduction was identified after noise education and using "noise lights" during lessons. Children and their teachers are equally exposed to high sound levels at school. Early sensitization to noise and the possible installation of sound-absorbing materials can be important means to prevent noise-associated hearing loss and mental impairment.
Newborn infants detect cues of concurrent sound segregation.
Bendixen, Alexandra; Háden, Gábor P; Németh, Renáta; Farkas, Dávid; Török, Miklós; Winkler, István
2015-01-01
Separating concurrent sounds is fundamental for a veridical perception of one's auditory surroundings. Sound components that are harmonically related and start at the same time are usually grouped into a common perceptual object, whereas components that are not in harmonic relation or have different onset times are more likely to be perceived in terms of separate objects. Here we tested whether neonates are able to pick up the cues supporting this sound organization principle. We presented newborn infants with a series of complex tones with their harmonics in tune (creating the percept of a unitary sound object) and with manipulated variants, which gave the impression of two concurrently active sound sources. The manipulated variant had either one mistuned partial (single-cue condition) or the onset of this mistuned partial was also delayed (double-cue condition). Tuned and manipulated sounds were presented in random order with equal probabilities. Recording the neonates' electroencephalographic responses allowed us to evaluate their processing of the sounds. Results show that, in both conditions, mistuned sounds elicited a negative displacement of the event-related potential (ERP) relative to tuned sounds from 360 to 400 ms after sound onset. The mistuning-related ERP component resembles the object-related negativity (ORN) component in adults, which is associated with concurrent sound segregation. Delayed onset additionally led to a negative displacement from 160 to 200 ms, which was probably more related to the physical parameters of the sounds than to their perceptual segregation. The elicitation of an ORN-like response in newborn infants suggests that neonates possess the basic capabilities of segregating concurrent sounds by detecting inharmonic relations between the co-occurring sounds. © 2015 S. Karger AG, Basel.
Acceptability of VTOL aircraft noise determined by absolute subjective testing
NASA Technical Reports Server (NTRS)
Sternfeld, H., Jr.; Hinterkeuser, E. G.; Hackman, R. B.; Davis, J.
1972-01-01
A program was conducted during which test subjects evaluated the simulated sounds of a helicopter, a tilt wing aircraft, and a 15 second, 90 PNdB (indoors) turbojet aircraft used as reference. Over 20,000 evaluations were made while the test subjects were engaged in work and leisure activities. The effects of level, exposure time, distance and aircraft design on subjective acceptability were evaluated. Some of the important conclusions are: (1) To be judged equal in annoyance to the reference jet sound, the helicopter and tilt wing sounds must be 4 to 5 PNdB lower when lasting 15 seconds in duration. (2) To be judged significantly more acceptable than the reference jet sound, the helicopter sound must be 10 PNdB lower when lasting 15 seconds in duration. (3) To be judged significantly more acceptable than the reference jet sound, the tilt wing sound must be 12 PNdB lower when lasting 15 seconds in duration. (4) The relative effect of changing the duration of a sound upon its subjectively rated annoyance diminishes with increasing duration. It varies from 2 PNdB per doubling of duration for intervals of 15 to 30 seconds, to 0.75 PNdB per doubling of duration for intervals of 120 to 240 seconds.
NASA Astrophysics Data System (ADS)
Miller, James D.
2003-10-01
A spiral model of pitch interrelates tone chroma, tone height, equal temperament scales, and a cochlear map. Donkin suggested in 1870 that the pitch of tones could be well represented by an equiangular spiral. More recently, the cylindrical helix has been popular for representing tone chroma and tone height. Here it is shown that tone chroma, tone height, and cochlear position can be conveniently related to tone frequency via a planar spiral. For this ``equal-temperament spiral,'' (ET Spiral) tone chroma is conceived as a circular array with semitones at 30° intervals. The frequency of sound on the cent scale (re 16.351 Hz) is represented by the radius of the spiral defined by r=(1200/2π)θr, where θr is in radians. By these definitions, one revolution represents one octave, 1200 cents, 30° represents a semitone, the radius relates θ to cents in accordance with equal temperament (ET) tuning, and the arclength of the spiral matches the mapping of sound frequency to the basilar membrane. Thus, the ET Spiral gives tone chroma as θ, tone height as the cent scale, and the cochlear map as the arclength. The possible implications and directions for further work are discussed.
Supersonic propeller noise in a uniform flow
NASA Technical Reports Server (NTRS)
Jou, Wen-Huei
1989-01-01
The sound field produced by a supersonic propeller operating in a uniform flow is investigated. The main interest is the effect of the finite forward flight speed on the directivity of the sound field as seen by an observer on the aircraft. It is found that there are cones of silence on the axis of the propeller. The semiapex angles on these cones are equal fore and aft of the propeller plane, and depend on the tip Mach number only. The Fourier coefficients of the acoustic pressure contain the Doppler amplification factor. The sound field weakens in the upstream direction and strengthen downstream. Kinematic considerations of the emitted Mach waves not only confirm these results, but also provide physical insight into the sound generation mechanism. The predicted zone of silence and the Doppler amplification factor are compared to the theoretical prediction of shock wave formation and the flight test of the SR3 propeller.
Code of Federal Regulations, 2010 CFR
2010-01-01
... a level that is sufficient to ensure the continued financial viability of the Enterprise and that equals or exceeds the minimum capital requirement contained in this subpart A. ... AND SOUNDNESS CAPITAL Minimum Capital § 1750.1 General. The regulation contained in this subpart A...
High intensity tone generation by axisymmetric ring cavities on training projectiles
NASA Technical Reports Server (NTRS)
Parthasarathy, S. P.; Cho, Y. I.; Back, L. H.
1984-01-01
An experimental investigation has been carried out on the production of high intensity tones by axisymmetric ring cavities. Maximum sound production occurs during a double resonance at Strouhal numbers which depend only on the local flow velocity independent of cavity location. Values of sound pressure of about 115 dB at 1 meter distance can be generated by axisymmetric ring cavities on projectiles moving at a relatively low flight speed equal to 65 m/s. Frequencies in the audible range up to several Kilo Hertz can be generated aeroacoustically.
A New Principle of Sound Frequency Analysis
NASA Technical Reports Server (NTRS)
Theodorsen, Theodore
1932-01-01
In connection with the study of aircraft and propeller noises, the National Advisory Committee for Aeronautics has developed an instrument for sound-frequency analysis which differs fundamentally from previous types, and which, owing to its simplicity of principle, construction, and operation, has proved to be of value in this investigation. The method is based on the well-known fact that the Ohmic loss in an electrical resistance is equal to the sum of the losses of the harmonic components of a complex wave, except for the case in which any two components approach or attain vectorial identity, in which case the Ohmic loss is increased by a definite amount. The principle of frequency analysis has been presented mathematically and a number of distinct advantages relative to previous methods have been pointed out. An automatic recording instrument embodying this principle is described in detail. It employs a beat-frequency oscillator as a source of variable frequency. A large number of experiments have verified the predicted superiority of the method. A number of representative records are presented.
Interpretation of magnetotelluric measurements over an electrically dispersive one-dimensional earth
DOE Office of Scientific and Technical Information (OSTI.GOV)
Patella, D.
1987-01-01
Frequency dispersion of electromagnetic parameters of earth materials has been widely documented in recent years. It is claimed that magnetotellurics (MT)may be significantly affected by dispersion. This paper studies the MT plane-wave interpretative problem for a one-dimensional earth characterized by the presence of dispersive layers. The theoretical properties of the MT field under the dispersion hypothesis, and the main features of the dispersion phenomenon are synthetically reviewed. The examination of previously published MT curve responses over some models of dispersive earth section shows that ambiguity can arise when interpreting MT data with no other source of information. Thus it maybemore » almost impossible to distinguish between the response of a dispersive section and an equally probable dispersion-free section. The dispersion magnetotelluric (DMT) method is proposed as a means to resolve the ambiguity. The DMT method is based on the execution, at the same site, of an MT sounding and of an always dispersion-free dc geoelectric deep sounding.« less
Christensen, Christian Bech; Christensen-Dalsgaard, Jakob; Brandt, Christian; Madsen, Peter Teglberg
2012-01-15
Snakes lack both an outer ear and a tympanic middle ear, which in most tetrapods provide impedance matching between the air and inner ear fluids and hence improve pressure hearing in air. Snakes would therefore be expected to have very poor pressure hearing and generally be insensitive to airborne sound, whereas the connection of the middle ear bone to the jaw bones in snakes should confer acute sensitivity to substrate vibrations. Some studies have nevertheless claimed that snakes are quite sensitive to both vibration and sound pressure. Here we test the two hypotheses that: (1) snakes are sensitive to sound pressure and (2) snakes are sensitive to vibrations, but cannot hear the sound pressure per se. Vibration and sound-pressure sensitivities were quantified by measuring brainstem evoked potentials in 11 royal pythons, Python regius. Vibrograms and audiograms showed greatest sensitivity at low frequencies of 80-160 Hz, with sensitivities of -54 dB re. 1 m s(-2) and 78 dB re. 20 μPa, respectively. To investigate whether pythons detect sound pressure or sound-induced head vibrations, we measured the sound-induced head vibrations in three dimensions when snakes were exposed to sound pressure at threshold levels. In general, head vibrations induced by threshold-level sound pressure were equal to or greater than those induced by threshold-level vibrations, and therefore sound-pressure sensitivity can be explained by sound-induced head vibration. From this we conclude that pythons, and possibly all snakes, lost effective pressure hearing with the complete reduction of a functional outer and middle ear, but have an acute vibration sensitivity that may be used for communication and detection of predators and prey.
Two-tone suppression in the cricket, Eunemobius carolinus (Gryllidae, Nemobiinae)
NASA Astrophysics Data System (ADS)
Farris, Hamilton E.; Hoy, Ronald R.
2002-03-01
Sounds with frequencies >15 kHz elicit an acoustic startle response (ASR) in flying crickets (Eunemobius carolinus). Although frequencies <15 kHz do not elicit the ASR when presented alone, when presented with ultrasound (40 kHz), low-frequency stimuli suppress the ultrasound-induced startle. Thus, using methods similar to those in masking experiments, we used two-tone suppression to assay sensitivity to frequencies in the audio band. Startle suppression was tuned to frequencies near 5 kHz, the frequency range of male calling songs. Similar to equal loudness contours measured in humans, however, equal suppression contours were not parallel, as the equivalent rectangular bandwidth of suppression tuning changed with increases in ultrasound intensity. Temporal integration of suppressor stimuli was measured using nonsimultaneous presentations of 5-ms pulses of 6 and 40 kHz. We found that no suppression occurs when the suppressing tone is >2 ms after and >5 ms before the ultrasound stimulus, suggesting that stimulus overlap is a requirement for suppression. When considered together with our finding that the intensity of low-frequency stimuli required for suppression is greater than that produced by singing males, the overlap requirement suggests that two-tone suppression functions to limit the ASR to sounds containing only ultrasound and not to broadband sounds that span the audio and ultrasound range.
Restoring speech perception with cochlear implants by spanning defective electrode contacts.
Frijns, Johan H M; Snel-Bongers, Jorien; Vellinga, Dirk; Schrage, Erik; Vanpoucke, Filiep J; Briaire, Jeroen J
2013-04-01
Even with six defective contacts, spanning can largely restore speech perception with the HiRes 120 speech processing strategy to the level supported by an intact electrode array. Moreover, the sound quality is not degraded. Previous studies have demonstrated reduced speech perception scores (SPS) with defective contacts in HiRes 120. This study investigated whether replacing defective contacts by spanning, i.e. current steering on non-adjacent contacts, is able to restore speech recognition to the level supported by an intact electrode array. Ten adult cochlear implant recipients (HiRes90K, HiFocus1J) with experience with HiRes 120 participated in this study. Three different defective electrode arrays were simulated (six separate defective contacts, three pairs or two triplets). The participants received three take-home strategies and were asked to evaluate the sound quality in five predefined listening conditions. After 3 weeks, SPS were evaluated with monosyllabic words in quiet and in speech-shaped background noise. The participants rated the sound quality equal for all take-home strategies. SPS with background noise were equal for all conditions tested. However, SPS in quiet (85% phonemes correct on average with the full array) decreased significantly with increasing spanning distance, with a 3% decrease for each spanned contact.
Yan, W Y; Li, L; Yang, Y G; Lin, X L; Wu, J Z
2016-08-01
We designed a computer-based respiratory sound analysis system to identify pediatric normal lung sound. To verify the validity of the computer-based respiratory sound analysis system. First we downloaded the standard lung sounds from the network database (website: http: //www.easyauscultation.com/lung-sounds-reference-guide) and recorded 3 samples of abnormal loud sound (rhonchi, wheeze and crackles) from three patients of The Department of Pediatrics, the First Affiliated Hospital of Xiamen University. We regarded such lung sounds as"reference lung sounds". The"test lung sounds"were recorded from 29 children form Kindergarten of Xiamen University. we recorded lung sound by portable electronic stethoscope and valid lung sounds were selected by manual identification. We introduced Mel-frequency cepstral coefficient (MFCC) to extract lung sound features and dynamic time warping (DTW) for signal classification. We had 39 standard lung sounds, recorded 58 test lung sounds. This computer-based respiratory sound analysis system was carried out in 58 lung sound recognition, correct identification of 52 times, error identification 6 times. Accuracy was 89.7%. Based on MFCC and DTW, our computer-based respiratory sound analysis system can effectively identify healthy lung sounds of children (accuracy can reach 89.7%), fully embodies the reliability of the lung sounds analysis system.
What is a melody? On the relationship between pitch and brightness of timbre
Cousineau, Marion; Carcagno, Samuele; Demany, Laurent; Pressnitzer, Daniel
2014-01-01
Previous studies showed that the perceptual processing of sound sequences is more efficient when the sounds vary in pitch than when they vary in loudness. We show here that sequences of sounds varying in brightness of timbre are processed with the same efficiency as pitch sequences. The sounds used consisted of two simultaneous pure tones one octave apart, and the listeners’ task was to make same/different judgments on pairs of sequences varying in length (one, two, or four sounds). In one condition, brightness of timbre was varied within the sequences by changing the relative level of the two pure tones. In other conditions, pitch was varied by changing fundamental frequency, or loudness was varied by changing the overall level. In all conditions, only two possible sounds could be used in a given sequence, and these two sounds were equally discriminable. When sequence length increased from one to four, discrimination performance decreased substantially for loudness sequences, but to a smaller extent for brightness sequences and pitch sequences. In the latter two conditions, sequence length had a similar effect on performance. These results suggest that the processes dedicated to pitch and brightness analysis, when probed with a sequence-discrimination task, share unexpected similarities. PMID:24478638
Sounds of Modified Flight Feathers Reliably Signal Danger in a Pigeon.
Murray, Trevor G; Zeil, Jochen; Magrath, Robert D
2017-11-20
In his book on sexual selection, Darwin [1] devoted equal space to non-vocal and vocal communication in birds. Since then, vocal communication has become a model for studies of neurobiology, learning, communication, evolution, and conservation [2, 3]. In contrast, non-vocal "instrumental music," as Darwin called it, has only recently become subject to sustained inquiry [4, 5]. In particular, outstanding work reveals how feathers, often highly modified, produce distinctive sounds [6-9], and suggests that these sounds have evolved at least 70 times, in many orders [10]. It remains to be shown, however, that such sounds are signals used in communication. Here we show that crested pigeons (Ochyphaps lophotes) signal alarm with specially modified wing feathers. We used video and feather-removal experiments to demonstrate that the highly modified 8 th primary wing feather (P8) produces a distinct note during each downstroke. The sound changes with wingbeat frequency, so that birds fleeing danger produce wing sounds with a higher tempo. Critically, a playback experiment revealed that only if P8 is present does the sound of escape flight signal danger. Our results therefore indicate, nearly 150 years after Darwin's book, that modified feathers can be used for non-vocal communication, and they reveal an intrinsically reliable alarm signal. Copyright © 2017 Elsevier Ltd. All rights reserved.
Using electronic storybooks to support word learning in children with severe language impairments.
Smeets, Daisy J H; van Dijken, Marianne J; Bus, Adriana G
2014-01-01
Novel word learning is reported to be problematic for children with severe language impairments (SLI). In this study, we tested electronic storybooks as a tool to support vocabulary acquisition in SLI children. In Experiment 1, 29 kindergarten SLI children heard four e-books each four times: (a) two stories were presented as video books with motion pictures, music, and sounds, and (b) two stories included only static illustrations without music or sounds. Two other stories served as the control condition. Both static and video books were effective in increasing knowledge of unknown words, but static books were most effective. Experiment 2 was designed to examine which elements in video books interfere with word learning: video images or music or sounds. A total of 23 kindergarten SLI children heard 8 storybooks each four times: (a) two static stories without music or sounds, (b) two static stories with music or sounds, (c) two video stories without music or sounds, and (d) two video books with music or sounds. Video images and static illustrations were equally effective, but the presence of music or sounds moderated word learning. In children with severe SLI, background music interfered with learning. Problems with speech perception in noisy conditions may be an underlying factor of SLI and should be considered in selecting teaching aids and learning environments. © Hammill Institute on Disabilities 2012.
Murgia, Mauro; Pili, Roberta; Corona, Federica; Sors, Fabrizio; Agostini, Tiziano A; Bernardis, Paolo; Casula, Carlo; Cossu, Giovanni; Guicciardi, Marco; Pau, Massimiliano
2018-01-01
The use of rhythmic auditory stimulation (RAS) has been proven useful in the management of gait disturbances associated with Parkinson's disease (PD). Typically, the RAS consists of metronome or music-based sounds (artificial RAS), while ecological footstep sounds (ecological RAS) have never been used for rehabilitation programs. The aim of this study was to compare the effects of a rehabilitation program integrated either with ecological or with artificial RAS. An observer-blind, randomized controlled trial was conducted to investigate the effects of 5 weeks of supervised rehabilitation integrated with RAS. Thirty-eight individuals affected by PD were randomly assigned to one of the two conditions (ecological vs. artificial RAS); thirty-two of them (age 68.2 ± 10.5, Hoehn and Yahr 1.5-3) concluded all phases of the study. Spatio-temporal parameters of gait and clinical variables were assessed before the rehabilitation period, at its end, and after a 3-month follow-up. Thirty-two participants were analyzed. The results revealed that both groups improved in the majority of biomechanical and clinical measures, independently of the type of sound. Moreover, exploratory analyses for separate groups were conducted, revealing improvements on spatio-temporal parameters only in the ecological RAS group. Overall, our results suggest that ecological RAS is equally effective compared to artificial RAS. Future studies should further investigate the role of ecological RAS, on the basis of information revealed by our exploratory analyses. Theoretical, methodological, and practical issues concerning the implementation of ecological sounds in the rehabilitation of PD patients are discussed. www.ClinicalTrials.gov, identifier NCT03228888.
Franzoni, Linda P; Elliott, Christopher M
2003-10-01
Experiments were performed on an elongated rectangular acoustic enclosure with different levels of absorptive material placed on side walls and an end wall. The acoustic source was a broadband high-frequency sound from a loudspeaker flush-mounted to an end wall of the enclosure. Measurements of sound-pressure levels were averaged in cross sections of the enclosure and then compared to theoretical results. Discrepancies between the experimental results and theoretical predictions that treated all incidence angles as equally probable led to the development of an angle-by-angle approach. The new approach agrees well with the experimentally obtained values. In addition, treating the absorptive material as bulk reacting rather than point reacting was found to significantly change the theoretical value for the absorption coefficient and to improve agreement with experiment. The new theory refines an earlier theory based on power conservation and locally diffuse assumptions. Furthermore, the new theory includes both the angle of incidence effects on the resistive and reactive properties of the absorptive material, and the effects of angle filtering, i.e., that reflecting waves associated with shallow angles become relatively stronger than those associated with steep angles as a function of distance from the source.
NASA Astrophysics Data System (ADS)
Araya, Mussie K.; Brownell, William E.
2015-12-01
Hearing requires precise detection and coding of acoustic signals by the inner ear and equally precise communication of the information through the auditory brainstem. A membrane based motor in the outer hair cell lateral wall contributes to the transformation of sound into a precise neural code. Structural, molecular and energetic similarities between the outer hair cell and auditory brainstem neurons suggest that a similar membrane based motor may contribute to signal processing in the auditory CNS. Cooperative activation of voltage gated ion channels enhances neuronal temporal processing and increases the upper frequency limit for phase locking. We explore the possibility that membrane mechanics contribute to ion channel cooperativity as a consequence of the nearly instantaneous speed of electromechanical signaling and the fact that membrane composition and mechanics modulate ion channel function.
Speech transformations based on a sinusoidal representation
NASA Astrophysics Data System (ADS)
Quatieri, T. E.; McAulay, R. J.
1986-05-01
A new speech analysis/synthesis technique is presented which provides the basis for a general class of speech transformation including time-scale modification, frequency scaling, and pitch modification. These modifications can be performed with a time-varying change, permitting continuous adjustment of a speaker's fundamental frequency and rate of articulation. The method is based on a sinusoidal representation of the speech production mechanism that has been shown to produce synthetic speech that preserves the waveform shape and is essentially perceptually indistinguishable from the original. Although the analysis/synthesis system originally was designed for single-speaker signals, it is equally capable of recovering and modifying nonspeech signals such as music; multiple speakers, marine biologic sounds, and speakers in the presence of interferences such as noise and musical backgrounds.
An aid in choosing the right tree to leave
C. Allen Bickford
1953-01-01
The successful forest manager has a number of well-stocked stands containing trees that are sound, useful, and fast-growing. Each stand and each tree is increasing in value fast enough that its rate of earning equals or exceeds the earnings from comparable investments.
Achieving perceptually-accurate aural telepresence
NASA Astrophysics Data System (ADS)
Henderson, Paul D.
Immersive multimedia requires not only realistic visual imagery but also a perceptually-accurate aural experience. A sound field may be presented simultaneously to a listener via a loudspeaker rendering system using the direct sound from acoustic sources as well as a simulation or "auralization" of room acoustics. Beginning with classical Wave-Field Synthesis (WFS), improvements are made to correct for asymmetries in loudspeaker array geometry. Presented is a new Spatially-Equalized WFS (SE-WFS) technique to maintain the energy-time balance of a simulated room by equalizing the reproduced spectrum at the listener for a distribution of possible source angles. Each reproduced source or reflection is filtered according to its incidence angle to the listener. An SE-WFS loudspeaker array of arbitrary geometry reproduces the sound field of a room with correct spectral and temporal balance, compared with classically-processed WFS systems. Localization accuracy of human listeners in SE-WFS sound fields is quantified by psychoacoustical testing. At a loudspeaker spacing of 0.17 m (equivalent to an aliasing cutoff frequency of 1 kHz), SE-WFS exhibits a localization blur of 3 degrees, nearly equal to real point sources. Increasing the loudspeaker spacing to 0.68 m (for a cutoff frequency of 170 Hz) results in a blur of less than 5 degrees. In contrast, stereophonic reproduction is less accurate with a blur of 7 degrees. The ventriloquist effect is psychometrically investigated to determine the effect of an intentional directional incongruence between audio and video stimuli. Subjects were presented with prerecorded full-spectrum speech and motion video of a talker's head as well as broadband noise bursts with a static image. The video image was displaced from the audio stimulus in azimuth by varying amounts, and the perceived auditory location measured. A strong bias was detectable for small angular discrepancies between audio and video stimuli for separations of less than 8 degrees for speech and less than 4 degrees with a pink noise burst. The results allow for the density of WFS systems to be selected from the required localization accuracy. Also, by exploiting the ventriloquist effect, the angular resolution of an audio rendering may be reduced when combined with spatially-accurate video.
Kok, Annebelle C M; Engelberts, J Pamela; Kastelein, Ronald A; Helder-Hoek, Lean; Van de Voorde, Shirley; Visser, Fleur; Slabbekoorn, Hans
2018-02-01
The continuing rise in underwater sound levels in the oceans leads to disturbance of marine life. It is thought that one of the main impacts of sound exposure is the alteration of foraging behaviour of marine species, for example by deterring animals from a prey location, or by distracting them while they are trying to catch prey. So far, only limited knowledge is available on both mechanisms in the same species. The harbour porpoise (Phocoena phocoena) is a relatively small marine mammal that could quickly suffer fitness consequences from a reduction of foraging success. To investigate effects of anthropogenic sound on their foraging efficiency, we tested whether experimentally elevated sound levels would deter two captive harbour porpoises from a noisy pool into a quiet pool (Experiment 1) and reduce their prey-search performance, measured as prey-search time in the noisy pool (Experiment 2). Furthermore, we tested the influence of the temporal structure and amplitude of the sound on the avoidance response of both animals. Both individuals avoided the pool with elevated sound levels, but they did not show a change in search time for prey when trying to find a fish hidden in one of three cages. The combination of temporal structure and SPL caused variable patterns. When the sound was intermittent, increased SPL caused increased avoidance times. When the sound was continuous, avoidance was equal for all SPLs above a threshold of 100 dB re 1 μPa. Hence, we found no evidence for an effect of sound exposure on search efficiency, but sounds of different temporal patterns did cause spatial avoidance with distinct dose-response patterns. Copyright © 2017 Elsevier Ltd. All rights reserved.
Sound reflection by a resonator array in a multimode cylindrical waveguide
NASA Astrophysics Data System (ADS)
Lapin, A. D.
2012-09-01
The paper considers the problem of scattering of the mth symmetric mode by an array of Q rings of identical, closely located Helmholtz resonators joined by necks to the walls of a wide circular pipe. The distance between rings is equal to half the wavelength of this mode at frequency ω, equal or close to the eigen-frequency of the resonator ring with allowance for the connected mass and interaction of neighboring rings via inhomogeneous modes. The coefficient of reflection of the mth mode from this grating array is calculated.
12 CFR 1207.21 - Equal opportunity in employment and contracting.
Code of Federal Regulations, 2012 CFR
2012-01-01
... MINORITY AND WOMEN INCLUSION Minority and Women Inclusion and Diversity at Regulated Entities and the... maximum extent possible in balance with financially safe and sound business practices, the inclusion and... direction of the officer immediately responsible for directing the Office of Minority and Women Inclusion...
12 CFR 1207.21 - Equal opportunity in employment and contracting.
Code of Federal Regulations, 2013 CFR
2013-01-01
... MINORITY AND WOMEN INCLUSION Minority and Women Inclusion and Diversity at Regulated Entities and the... maximum extent possible in balance with financially safe and sound business practices, the inclusion and... direction of the officer immediately responsible for directing the Office of Minority and Women Inclusion...
12 CFR 1207.21 - Equal opportunity in employment and contracting.
Code of Federal Regulations, 2014 CFR
2014-01-01
... OPERATIONS MINORITY AND WOMEN INCLUSION Minority and Women Inclusion and Diversity at Regulated Entities and... maximum extent possible in balance with financially safe and sound business practices, the inclusion and... direction of the officer immediately responsible for directing the Office of Minority and Women Inclusion...
On remote sounding of the upper atmosphere of Venus
NASA Technical Reports Server (NTRS)
Houghton, J. T.; Taylor, F. W.
1975-01-01
Some of the possibilities for remote sensing of the upper atmosphere of Venus from an orbiting spacecraft are studied quantitatively. Temperature sounding over a wide vertical range, from the main cloud top near 60 km altitude to the nanobar level near 160 km, is shown to be feasible. Techniques which deconvolve the cloud structure from the temperature profile measurements are examined. Humidity measurements by simple radiometry are feasible for column abundances greater than or equal to 10 precipitable micrometers. The information content of limb radiance measurements, in different wavelengths and for various viewing geometries, is also analyzed.
NASA Astrophysics Data System (ADS)
Houjoh, Haruo
1992-12-01
One specific feature of the aerodynamic sound produced at the face end region is that the radiation becomes equally weak by filling root spaces as by shortening the center distance. However, one can easily expect that such actions make the air flow faster, and consequently make the sound louder. This paper attempts to reveal the reason for such a feature. First, air flow induced by the pumping action of the gear pair was analyzed regarding a series of root spaces as volume varying cavities which have channels to adjacent cavities as well as the exit/inlet at the face ends. The numerical analysis was verified by the hot wire anemometer measurement. Next, from the obtained flow response, the sound source was estimated to be a combination of symmetrically distributed simple sources. Taking the effect of either the center distance or root filling into consideration, it is shown that the simplified model can explain such a feature rationally.
Auditory Discrimination of Frequency Ratios: The Octave Singularity
ERIC Educational Resources Information Center
Bonnard, Damien; Micheyl, Christophe; Semal, Catherine; Dauman, Rene; Demany, Laurent
2013-01-01
Sensitivity to frequency ratios is essential for the perceptual processing of complex sounds and the appreciation of music. This study assessed the effect of ratio simplicity on ratio discrimination for pure tones presented either simultaneously or sequentially. Each stimulus consisted of four 100-ms pure tones, equally spaced in terms of…
The Delirious Spectator: Opening Spaces in Film Studies
ERIC Educational Resources Information Center
Sommer, Paul
2017-01-01
The 2016 "Garth Boomer Address" considers the teaching of film. It challenges the orthodoxy that calls for a predominantly visual analysis, arguing that editing, sound and a fresh look at the script are equally important. This invites an understanding of characterisation and narrative in terms of the creation of "wholes" and…
Michaud, Mark; Leong, Thomas; Swiergon, Piotr; Juliano, Pablo; Knoerzer, Kai
2015-09-01
This work validated, in a higher frequency range, the theoretical predictions made by Boyle around 1930, which state that the optimal transmission of sound pressure through a metal plate occurs when the plate thickness equals a multiple of half the wavelength of the sound wave. Several reactor design parameters influencing the transmission of high frequency ultrasonic waves through a stainless steel plate were examined. The transmission properties of steel plates of various thicknesses (1-7 mm) were studied for frequencies ranging from 400 kHz to 2 MHz and at different distances between plates and transducers. It was shown that transmission of sound pressure through a steel plate showed high dependence of the thickness of the plate to the frequency of the sound wave (thickness ratio). Maximum sound pressure transmission of ∼ 60% of the incident pressure was observed when the ratio of the plate thickness to the applied frequency was a multiple of a half wavelength (2 MHz, 6mm stainless steel plate). In contrast, minimal sound pressure transmission (∼ 10-20%) was measured for thickness ratios that were not a multiple of a half wavelength. Furthermore, the attenuation of the sound pressure in the transmission region was also investigated. As expected, it was confirmed that higher frequencies have more pronounced sound pressure attenuation than lower frequencies. The spatial distribution of the sound pressure transmitted through the plate characterized by sonochemiluminescence measurements using luminol emission, supports the validity of the pressure measurements in this study. Copyright © 2015 Elsevier B.V. All rights reserved.
A system to simulate and reproduce audio-visual environments for spatial hearing research.
Seeber, Bernhard U; Kerber, Stefan; Hafter, Ervin R
2010-02-01
The article reports the experience gained from two implementations of the "Simulated Open-Field Environment" (SOFE), a setup that allows sounds to be played at calibrated levels over a wide frequency range from multiple loudspeakers in an anechoic chamber. Playing sounds from loudspeakers in the free-field has the advantage that each participant listens with their own ears, and individual characteristics of the ears are captured in the sound they hear. This makes an easy and accurate comparison between various listeners with and without hearing devices possible. The SOFE uses custom calibration software to assure individual equalization of each loudspeaker. Room simulation software creates the spatio-temporal reflection pattern of sound sources in rooms which is played via the SOFE loudspeakers. The sound playback system is complemented by a video projection facility which can be used to collect or give feedback or to study auditory-visual interaction. The article discusses acoustical and technical requirements for accurate sound playback against the specific needs in hearing research. An introduction to software concepts is given which allow easy, high-level control of the setup and thus fast experimental development, turning the SOFE into a "Swiss army knife" tool for auditory, spatial hearing and audio-visual research. Crown Copyright 2009. Published by Elsevier B.V. All rights reserved.
A System to Simulate and Reproduce Audio-Visual Environments for Spatial Hearing Research
Seeber, Bernhard U.; Kerber, Stefan; Hafter, Ervin R.
2009-01-01
The article reports the experience gained from two implementations of the “Simulated Open-Field Environment” (SOFE), a setup that allows sounds to be played at calibrated levels over a wide frequency range from multiple loudspeakers in an anechoic chamber. Playing sounds from loudspeakers in the free-field has the advantage that each participant listens with their own ears, and individual characteristics of the ears are captured in the sound they hear. This makes an easy and accurate comparison between various listeners with and without hearing devices possible. The SOFE uses custom calibration software to assure individual equalization of each loudspeaker. Room simulation software creates the spatio-temporal reflection pattern of sound sources in rooms which is played via the SOFE loudspeakers. The sound playback system is complemented by a video projection facility which can be used to collect or give feedback or to study auditory-visual interaction. The article discusses acoustical and technical requirements for accurate sound playback against the specific needs in hearing research. An introduction to software concepts is given which allow easy, high-level control of the setup and thus fast experimental development, turning the SOFE into a “Swiss army knife” tool for auditory, spatial hearing and audio-visual research. PMID:19909802
Cosmological perturbations in mimetic Horndeski gravity
NASA Astrophysics Data System (ADS)
Arroja, Frederico; Bartolo, Nicola; Karmakar, Purnendu; Matarrese, Sabino
2016-04-01
We study linear scalar perturbations around a flat FLRW background in mimetic Horndeski gravity. In the absence of matter, we show that the Newtonian potential satisfies a second-order differential equation with no spatial derivatives. This implies that the sound speed for scalar perturbations is exactly zero on this background. We also show that in mimetic G3 theories the sound speed is equally zero. We obtain the equation of motion for the comoving curvature perturbation (first order differential equation) and solve it to find that the comoving curvature perturbation is constant on all scales in mimetic Horndeski gravity. We find solutions for the Newtonian potential evolution equation in two simple models. Finally we show that the sound speed is zero on all backgrounds and therefore the system does not have any wave-like scalar degrees of freedom.
Annoyance caused by the sounds of a magnetic levitation train
NASA Astrophysics Data System (ADS)
Vos, Joos
2004-04-01
In a laboratory study, the annoyance caused by the passby sounds from a magnetic levitation (maglev) train was investigated. The listeners were presented with various sound fragments. The task of the listeners was to respond after each presentation to the question: ``How annoying would you find the sound in the preceding period if you were exposed to it at home on a regular basis?'' The independent variables were (a) the driving speed of the maglev train (varying from 100 to 400 km/h), (b) the outdoor A-weighted sound exposure level (ASEL) of the passbys (varying from 65 to 90 dB), and (c) the simulated outdoor-to-indoor reduction in sound level (windows open or windows closed). As references to the passby sounds from the maglev train (type Transrapid 08), sounds from road traffic (passenger cars and trucks) and more conventional railway (intercity trains) were included for rating also. Four important results were obtained. Provided that the outdoor ASELs were the same, (1) the annoyance was independent of the driving speed of the maglev train, (2) the annoyance caused by the maglev train was considerably higher than that caused by the intercity train, (3) the annoyance caused by the maglev train was hardly different from that caused by road traffic, and (4) the results (1)-(3) held true both for open or closed windows. On the basis of the present results, it might be expected that the sounds are equally annoying if the ASELs of the maglev-train passbys are at least 5 dB lower than those of the intercity train passbys. Consequently, the results of the present experiment do not support application of a railway bonus to the maglev-train sounds.
Annoyance caused by the sounds of a magnetic levitation train.
Vos, Joos
2004-04-01
In a laboratory study, the annoyance caused by the passby sounds from a magnetic levitation (maglev) train was investigated. The listeners were presented with various sound fragments. The task of the listeners was to respond after each presentation to the question: "How annoying would you find the sound in the preceding period if you were exposed to it at home on a regular basis?" The independent variables were (a) the driving speed of the maglev train (varying from 100 to 400 km/h), (b) the outdoor A-weighted sound exposure level (ASEL) of the passbys (varying from 65 to 90 dB), and (c) the simulated outdoor-to-indoor reduction in sound level (windows open or windows closed). As references to the passby sounds from the maglev train (type Transrapid 08), sounds from road traffic (passenger cars and trucks) and more conventional railway (intercity trains) were included for rating also. Four important results were obtained. Provided that the outdoor ASELs were the same, (1) the annoyance was independent of the driving speed of the maglev train, (2) the annoyance caused by the maglev train was considerably higher than that caused by the intercity train, (3) the annoyance caused by the maglev train was hardly different from that caused by road traffic, and (4) the results (1)-(3) held true both for open or closed windows. On the basis of the present results, it might be expected that the sounds are equally annoying if the ASELs of the maglev-train passbys are at least 5 dB lower than those of the intercity train passbys. Consequently, the results of the present experiment do not support application of a railway bonus to the maglev-train sounds.
Loebach, Jeremy L; Pisoni, David B; Svirsky, Mario A
2009-12-01
The objective of this study was to assess whether training on speech processed with an eight-channel noise vocoder to simulate the output of a cochlear implant would produce transfer of auditory perceptual learning to the recognition of nonspeech environmental sounds, the identification of speaker gender, and the discrimination of talkers by voice. Twenty-four normal-hearing subjects were trained to transcribe meaningful English sentences processed with a noise vocoder simulation of a cochlear implant. An additional 24 subjects served as an untrained control group and transcribed the same sentences in their unprocessed form. All subjects completed pre- and post-test sessions in which they transcribed vocoded sentences to provide an assessment of training efficacy. Transfer of perceptual learning was assessed using a series of closed set, nonlinguistic tasks: subjects identified talker gender, discriminated the identity of pairs of talkers, and identified ecologically significant environmental sounds from a closed set of alternatives. Although both groups of subjects showed significant pre- to post-test improvements, subjects who transcribed vocoded sentences during training performed significantly better at post-test than those in the control group. Both groups performed equally well on gender identification and talker discrimination. Subjects who received explicit training on the vocoded sentences, however, performed significantly better on environmental sound identification than the untrained subjects. Moreover, across both groups, pre-test speech performance and, to a higher degree, post-test speech performance, were significantly correlated with environmental sound identification. For both groups, environmental sounds that were characterized as having more salient temporal information were identified more often than environmental sounds that were characterized as having more salient spectral information. Listeners trained to identify noise-vocoded sentences showed evidence of transfer of perceptual learning to the identification of environmental sounds. In addition, the correlation between environmental sound identification and sentence transcription indicates that subjects who were better able to use the degraded acoustic information to identify the environmental sounds were also better able to transcribe the linguistic content of novel sentences. Both trained and untrained groups performed equally well ( approximately 75% correct) on the gender-identification task, indicating that training did not have an effect on the ability to identify the gender of talkers. Although better than chance, performance on the talker discrimination task was poor overall ( approximately 55%), suggesting that either explicit training is required to discriminate talkers' voices reliably or that additional information (perhaps spectral in nature) not present in the vocoded speech is required to excel in such tasks. Taken together, the results suggest that although transfer of auditory perceptual learning with spectrally degraded speech does occur, explicit task-specific training may be necessary for tasks that cannot rely on temporal information alone.
Loebach, Jeremy L.; Pisoni, David B.; Svirsky, Mario A.
2009-01-01
Objective The objective of this study was to assess whether training on speech processed with an 8-channel noise vocoder to simulate the output of a cochlear implant would produce transfer of auditory perceptual learning to the recognition of non-speech environmental sounds, the identification of speaker gender, and the discrimination of talkers by voice. Design Twenty-four normal hearing subjects were trained to transcribe meaningful English sentences processed with a noise vocoder simulation of a cochlear implant. An additional twenty-four subjects served as an untrained control group and transcribed the same sentences in their unprocessed form. All subjects completed pre- and posttest sessions in which they transcribed vocoded sentences to provide an assessment of training efficacy. Transfer of perceptual learning was assessed using a series of closed-set, nonlinguistic tasks: subjects identified talker gender, discriminated the identity of pairs of talkers, and identified ecologically significant environmental sounds from a closed set of alternatives. Results Although both groups of subjects showed significant pre- to posttest improvements, subjects who transcribed vocoded sentences during training performed significantly better at posttest than subjects in the control group. Both groups performed equally well on gender identification and talker discrimination. Subjects who received explicit training on the vocoded sentences, however, performed significantly better on environmental sound identification than the untrained subjects. Moreover, across both groups, pretest speech performance, and to a higher degree posttest speech performance, were significantly correlated with environmental sound identification. For both groups, environmental sounds that were characterized as having more salient temporal information were identified more often than environmental sounds that were characterized as having more salient spectral information. Conclusions Listeners trained to identify noise-vocoded sentences showed evidence of transfer of perceptual learning to the identification of environmental sounds. In addition, the correlation between environmental sound identification and sentence transcription indicates that subjects who were better able to utilize the degraded acoustic information to identify the environmental sounds were also better able to transcribe the linguistic content of novel sentences. Both trained and untrained groups performed equally well (~75% correct) on the gender identification task, indicating that training did not have an effect on the ability to identify the gender of talkers. Although better than chance, performance on the talker discrimination task was poor overall (~55%), suggesting that either explicit training is required to reliably discriminate talkers’ voices, or that additional information (perhaps spectral in nature) not present in the vocoded speech is required to excel in such tasks. Taken together, the results suggest that while transfer of auditory perceptual learning with spectrally degraded speech does occur, explicit task-specific training may be necessary for tasks that cannot rely on temporal information alone. PMID:19773659
The Radio Plasma Imager Investigation on the IMAGE Spacecraft
NASA Technical Reports Server (NTRS)
Reinisch, Bodo W.; Haines, D. M.; Bibl, K.; Cheney, G.; Galkin, I. A.; Huang, X.; Myers, S. H.; Sales, G. S.; Benson, R. F.; Fung, S. F.
1999-01-01
Radio plasma imaging uses total reflection of electromagnetic waves from plasmas whose plasma frequencies equal the radio sounding frequency and whose electron density gradients are parallel to the wave normals. The Radio Plasma Imager (RPI) has two orthogonal 500-m long dipole antennas in the spin plane for near omni-directional transmission. The third antenna is a 20-m dipole. Echoes from the magnetopause, plasmasphere and cusp will be received with three orthogonal antennas, allowing the determination of their angle-of-arrival. Thus it will be possible to create image fragments of the reflecting density structures. The instrument can execute a large variety of programmable measuring programs operating at frequencies between 3 kHz and 3 MHz. Tuning of the transmit antennas provides optimum power transfer from the 10 W transmitter to the antennas. The instrument can operate in three active sounding modes: (1) remote sounding to probe magnetospheric boundaries, (2) local (relaxation) sounding to probe the local plasma, and (3) whistler stimulation sounding. In addition, there is a passive mode to record natural emissions, and to determine the local electron density and temperature by using a thermal noise spectroscopy technique.
NFL Films audio, video, and film production facilities
NASA Astrophysics Data System (ADS)
Berger, Russ; Schrag, Richard C.; Ridings, Jason J.
2003-04-01
The new NFL Films 200,000 sq. ft. headquarters is home for the critically acclaimed film production that preserves the NFL's visual legacy week-to-week during the football season, and is also the technical plant that processes and archives football footage from the earliest recorded media to the current network broadcasts. No other company in the country shoots more film than NFL Films, and the inclusion of cutting-edge video and audio formats demands that their technical spaces continually integrate the latest in the ever-changing world of technology. This facility houses a staggering array of acoustically sensitive spaces where music and sound are equal partners with the visual medium. Over 90,000 sq. ft. of sound critical technical space is comprised of an array of sound stages, music scoring stages, audio control rooms, music writing rooms, recording studios, mixing theaters, video production control rooms, editing suites, and a screening theater. Every production control space in the building is designed to monitor and produce multi channel surround sound audio. An overview of the architectural and acoustical design challenges encountered for each sophisticated listening, recording, viewing, editing, and sound critical environment will be discussed.
Noise annoyance caused by magnetic levitation train passbys
NASA Astrophysics Data System (ADS)
Vos, Joos
2004-05-01
In a laboratory study, the annoyance caused by the passby sounds from a magnetic levitation (maglev) train was investigated. The outdoor A-weighted sound exposure level (ASEL) of the maglev sounds varied from 65 to 90 dB. The driving speed of the maglev train varied from 100 to 400 km/h. Four important results were obtained. Provided that the outdoor ASELs were the same, (1) the annoyance was independent of the driving speed of the maglev train, (2) the annoyance caused by the maglev train was considerably higher than that caused by intercity trains, (3) the annoyance caused by the maglev train was hardly different from that caused by road traffic (passenger cars and trucks), and (4) the results (1)-(3) held true both for open or closed windows. On the basis of the present results, it might be expected that the sounds are equally annoying if the ASELs of the maglev-train passbys are at least 5 dB lower than those of the intercity train passbys. Consequently, the results of the present experiment do not support application of a railway bonus to the maglev-train sounds. Issues for future research, such as exploring further contributions of nonacoustic factors, will be discussed.
System and technique for characterizing fluids using ultrasonic diffraction grating spectroscopy
Greenwood, Margaret S.
2005-04-12
A system for determining a property of a fluid based on ultrasonic diffraction grating spectroscopy includes a diffraction grating on a solid in contact with the fluid. An interrogation device delivers ultrasound through the solid and a captures a reflection spectrum from the diffraction grating. The reflection spectrum including a diffraction order equal to zero exhibits a peak whose location is used to determine speed of sound in the fluid. A separate measurement of the acoustic impedance is combined with the determined speed of sound to yield a measure of fluid density. A system for determining acoustic impedance includes an ultrasonic transducer on a first surface of a solid member, and an opposed second surface of the member is in contact with a fluid to be monitored. A longitudinal ultrasonic pulse is delivered through the solid member, and a multiplicity of pulse echoes caused by reflections of the ultrasonic pulse between the solid-fluid interface and the transducer-solid interface are detected. The decay rate of the detected echo amplitude as a function of echo number is used to determine acoustic impedance.
New numerical solutions of three-dimensional compressible hydrodynamic convection. [in stars
NASA Technical Reports Server (NTRS)
Hossain, Murshed; Mullan, D. J.
1990-01-01
Numerical solutions of three-dimensional compressible hydrodynamics (including sound waves) in a stratified medium with open boundaries are presented. Convergent/divergent points play a controlling role in the flows, which are dominated by a single frequency related to the mean sound crossing time. Superposed on these rapid compressive flows, slower eddy-like flows eventually create convective transport. The solutions contain small structures stacked on top of larger ones, with vertical scales equal to the local pressure scale heights, H sub p. Although convective transport starts later in the evolution, vertical scales of H sub p are apparently selected at much earlier times by nonlinear compressive effects.
Multimodal far-field acoustic radiation pattern: An approximate equation
NASA Technical Reports Server (NTRS)
Rice, E. J.
1977-01-01
The far-field sound radiation theory for a circular duct was studied for both single mode and multimodal inputs. The investigation was intended to develop a method to determine the acoustic power produced by turbofans as a function of mode cut-off ratio. With reasonable simplifying assumptions the single mode radiation pattern was shown to be reducible to a function of mode cut-off ratio only. With modal cut-off ratio as the dominant variable, multimodal radiation patterns can be reduced to a simple explicit expression. This approximate expression provides excellent agreement with an exact calculation of the sound radiation pattern using equal acoustic power per mode.
A Sound Education for All: Multicultural Issues in Music Education
ERIC Educational Resources Information Center
Johnson, Jr., Bob L.
2004-01-01
Establishing the legitimacy of the arts within the larger school curriculum is a defining issue in arts education. Within the context of this perennial challenge, this article examines two multicultural issues in music education: equal music education opportunity and the idiomatic hegemony of the Western classical tradition. Discussions of the…
The Bounds on Flexibility in Speech Perception
ERIC Educational Resources Information Center
Sjerps, Matthias J.; McQueen, James M.
2010-01-01
Dutch listeners were exposed to the English theta sound (as in "bath"), which replaced [f] in /f/-final Dutch words or, for another group, [s] in /s/-final words. A subsequent identity-priming task showed that participants had learned to interpret theta as, respectively, /f/ or /s/. Priming effects were equally strong when the exposure…
NASA Astrophysics Data System (ADS)
Jagadeesha, C. B.
2017-12-01
Even though friction stir welding was invented long back (1991) by TWI England, till now there has no method or procedure or approach developed, which helps to obtain quickly optimum or exact parameters yielding good or sound weld. An approach has developed in which an equation has been derived, by which approximate rpm can be obtained and by setting range of rpm ±100 or 50 rpm over approximate rpm and by setting welding speed equal to 60 mm/min or 50 mm/min one can conduct FSW experiment to reach optimum parameters; one can reach quickly to optimum parameters, i.e. desired rpm, and welding speed, which yield sound weld by the approach. This approach can be effectively used to obtain sound welds for all similar and dissimilar combinations of materials such as Steel, Al, Mg, Ti, etc.
NASA Astrophysics Data System (ADS)
Miner, Nadine Elizabeth
1998-09-01
This dissertation presents a new wavelet-based method for synthesizing perceptually convincing, dynamic sounds using parameterized sound models. The sound synthesis method is applicable to a variety of applications including Virtual Reality (VR), multi-media, entertainment, and the World Wide Web (WWW). A unique contribution of this research is the modeling of the stochastic, or non-pitched, sound components. This stochastic-based modeling approach leads to perceptually compelling sound synthesis. Two preliminary studies conducted provide data on multi-sensory interaction and audio-visual synchronization timing. These results contributed to the design of the new sound synthesis method. The method uses a four-phase development process, including analysis, parameterization, synthesis and validation, to create the wavelet-based sound models. A patent is pending for this dynamic sound synthesis method, which provides perceptually-realistic, real-time sound generation. This dissertation also presents a battery of perceptual experiments developed to verify the sound synthesis results. These experiments are applicable for validation of any sound synthesis technique.
Toward Inverse Control of Physics-Based Sound Synthesis
NASA Astrophysics Data System (ADS)
Pfalz, A.; Berdahl, E.
2017-05-01
Long Short-Term Memory networks (LSTMs) can be trained to realize inverse control of physics-based sound synthesizers. Physics-based sound synthesizers simulate the laws of physics to produce output sound according to input gesture signals. When a user's gestures are measured in real time, she or he can use them to control physics-based sound synthesizers, thereby creating simulated virtual instruments. An intriguing question is how to program a computer to learn to play such physics-based models. This work demonstrates that LSTMs can be trained to accomplish this inverse control task with four physics-based sound synthesizers.
Experimental demonstration of multiuser communication in deep water using time reversal.
Shimura, T; Ochi, H; Song, H C
2013-10-01
Multiuser communication is demonstrated using experimental data (450-550 Hz) collected in deep water, south of Japan. The multiple users are spatially distributed either in depth or range while a 114-m long, 20-element vertical array (i.e., base station) is deployed to around the sound channel axis (~1000 m). First, signals received separately from ranges of 150 km and 180 km at various depths are combined asynchronously to generate multiuser communication sequences for subsequent processing, achieving an aggregate data rate of 300 bits/s for up to three users. Adaptive time reversal is employed to separate collided packets at the base station, followed by a single channel decision feedback equalizer. Then it is demonstrated that two users separated by 3 km in range at ~1000 m depth can transmit information simultaneously to the base station at ~500 km range with an aggregate data rate of 200 bits/s.
Bidet-Caulet, Aurélie; Buchanan, Kelly G; Viswanath, Humsini; Black, Jessica; Scabini, Donatella; Bonnet-Brilhault, Frédérique; Knight, Robert T
2015-11-01
There is growing evidence that auditory selective attention operates via distinct facilitatory and inhibitory mechanisms enabling selective enhancement and suppression of sound processing, respectively. The lateral prefrontal cortex (LPFC) plays a crucial role in the top-down control of selective attention. However, whether the LPFC controls facilitatory, inhibitory, or both attentional mechanisms is unclear. Facilitatory and inhibitory mechanisms were assessed, in patients with LPFC damage, by comparing event-related potentials (ERPs) to attended and ignored sounds with ERPs to these same sounds when attention was equally distributed to all sounds. In control subjects, we observed 2 late frontally distributed ERP components: a transient facilitatory component occurring from 150 to 250 ms after sound onset; and an inhibitory component onsetting at 250 ms. Only the facilitatory component was affected in patients with LPFC damage: this component was absent when attending to sounds delivered in the ear contralateral to the lesion, with the most prominent decreases observed over the damaged brain regions. These findings have 2 important implications: (i) they provide evidence for functionally distinct facilitatory and inhibitory mechanisms supporting late auditory selective attention; (ii) they show that the LPFC is involved in the control of the facilitatory mechanisms of auditory attention. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
Relationships between early literacy and nonlinguistic rhythmic processes in kindergarteners.
Ozernov-Palchik, Ola; Wolf, Maryanne; Patel, Aniruddh D
2018-03-01
A growing number of studies report links between nonlinguistic rhythmic abilities and certain linguistic abilities, particularly phonological skills. The current study investigated the relationship between nonlinguistic rhythmic processing, phonological abilities, and early literacy abilities in kindergarteners. A distinctive aspect of the current work was the exploration of whether processing of different types of rhythmic patterns is differentially related to kindergarteners' phonological and reading-related abilities. Specifically, we examined the processing of metrical versus nonmetrical rhythmic patterns, that is, patterns capable of being subdivided into equal temporal intervals or not (Povel & Essens, 1985). This is an important comparison because most music involves metrical sequences, in which rhythm often has an underlying temporal grid of isochronous units. In contrast, nonmetrical sequences are arguably more typical to speech rhythm, which is temporally structured but does not involve an underlying grid of equal temporal units. A rhythm discrimination app with metrical and nonmetrical patterns was administered to 74 kindergarteners in conjunction with cognitive and preliteracy measures. Findings support a relationship among rhythm perception, phonological awareness, and letter-sound knowledge (an essential precursor of reading). A mediation analysis revealed that the association between rhythm perception and letter-sound knowledge is mediated through phonological awareness. Furthermore, metrical perception accounted for unique variance in letter-sound knowledge above all other language and cognitive measures. These results point to a unique role for temporal regularity processing in the association between musical rhythm and literacy in young children. Copyright © 2017 Elsevier Inc. All rights reserved.
A laboratory study of subjective annoyance response to sonic booms and aircraft flyovers
NASA Technical Reports Server (NTRS)
Leatherwood, Jack D.; Sullivan, Brenda M.
1994-01-01
Three experiments were conducted to determine subjective equivalence of aircraft subsonic flyover noise and sonic booms. Two of the experiments were conducted in a loudspeaker-driven sonic boom simulator, and the third in a large room containing conventional loudspeakers. The sound generation system of the boom simulator had a frequency response extending to very low frequencies (about 1 Hz) whereas the large room loudspeakers were limited to about 20 Hz. Subjective equivalence between booms and flyovers was quantified in terms of the difference between the noise level of a boom and that of a flyover when the two were judged equally annoying. Noise levels were quantified in terms of the following noise descriptors: Perceived Level (PL), Perceived Noise Level (PNL), C-weighted sound exposure level (SELC), and A-weighted sound exposure level (SELA). Results from the present study were compared, where possible, to similar results obtained in other studies. Results showed that noise level differences depended upon the descriptor used, specific boom and aircraft noise events being compared and, except for the PNL descriptor, varied between the simulator and large room. Comparison of noise level differences obtained in the present study with those of other studies indicated good agreement across studies only for the PNL and SELA descriptors. Comparison of the present results with assessments of community response to high-energy impulsive sounds made by Working Group 84 of the National Research Council's Committee on Hearing, Bioacoustics, and Biomechanics (CHABA) showed good agreement when boom/flyover noise level differences were based on SELA. However, noise level differences obtained by CHABA using SELA for aircraft flyovers and SELC for booms were not in agreement with results obtained in the present study.
Neves, A A; Silva, E J; Roter, J M; Belladona, F G; Alves, H D; Lopes, R T; Paciornik, S; De-Deus, G A
2015-11-01
To propose an automated image processing routine based on free software to quantify root canal preparation outcomes in pairs of sound and instrumented roots after micro-CT scanning procedures. Seven mesial roots of human mandibular molars with different canal configuration systems were studied: (i) Vertucci's type 1, (ii) Vertucci's type 2, (iii) two individual canals, (iv) Vertucci's type 6, canals (v) with and (vi) without debris, and (vii) canal with visible pulp calcification. All teeth were instrumented with the BioRaCe system and scanned in a Skyscan 1173 micro-CT before and after canal preparation. After reconstruction, the instrumented stack of images (IS) was registered against the preoperative sound stack of images (SS). Image processing included contrast equalization and noise filtering. Sound canal volumes were obtained by a minimum threshold. For the IS, a fixed conservative threshold was chosen as the best compromise between instrumented canal and dentine whilst avoiding debris, resulting in instrumented canal plus empty spaces. Arithmetic and logical operations between sound and instrumented stacks were used to identify debris. Noninstrumented dentine was calculated using a minimum threshold in the IS and subtracting from the SS and total debris. Removed dentine volume was obtained by subtracting SS from IS. Quantitative data on total debris present in the root canal space after instrumentation, noninstrumented areas and removed dentine volume were obtained for each test case, as well as three-dimensional volume renderings. After standardization of acquisition, reconstruction and image processing micro-CT images, a quantitative approach for calculation of root canal biomechanical outcomes was achieved using free software. © 2014 International Endodontic Journal. Published by John Wiley & Sons Ltd.
van Atteveldt, Nienke; Musacchia, Gabriella; Zion-Golumbic, Elana; Sehatpour, Pejman; Javitt, Daniel C.; Schroeder, Charles
2015-01-01
The brain’s fascinating ability to adapt its internal neural dynamics to the temporal structure of the sensory environment is becoming increasingly clear. It is thought to be metabolically beneficial to align ongoing oscillatory activity to the relevant inputs in a predictable stream, so that they will enter at optimal processing phases of the spontaneously occurring rhythmic excitability fluctuations. However, some contexts have a more predictable temporal structure than others. Here, we tested the hypothesis that the processing of rhythmic sounds is more efficient than the processing of irregularly timed sounds. To do this, we simultaneously measured functional magnetic resonance imaging (fMRI) and electro-encephalograms (EEG) while participants detected oddball target sounds in alternating blocks of rhythmic (e.g., with equal inter-stimulus intervals) or random (e.g., with randomly varied inter-stimulus intervals) tone sequences. Behaviorally, participants detected target sounds faster and more accurately when embedded in rhythmic streams. The fMRI response in the auditory cortex was stronger during random compared to random tone sequence processing. Simultaneously recorded N1 responses showed larger peak amplitudes and longer latencies for tones in the random (vs. the rhythmic) streams. These results reveal complementary evidence for more efficient neural and perceptual processing during temporally predictable sensory contexts. PMID:26579044
Open ear hearing aids in tinnitus therapy: An efficacy comparison with sound generators.
Parazzini, Marta; Del Bo, Luca; Jastreboff, Margaret; Tognola, Gabriella; Ravazzani, Paolo
2011-08-01
This study aimed to compare the effectiveness of tinnitus retraining therapy (TRT) with sound generators or with open ear hearing aids in the rehabilitation of tinnitus for a group of subjects who, according to Jastreboff categories, can be treated with both approaches to sound therapy (borderline of Category 1 and 2). This study was a prospective data collection with a parallel-group design which entailed that each subject was randomly assigned to one of the two treatments group: half of the subjects were fitted binaurally with sound generators, and the other half with open ear hearing aids. Both groups received the same educational counselling sessions. Ninety-one subjects passed the screening criteria and were enrolled into the study. Structured interviews, with a variety of measures evaluated through the use of visual-analog scales and the tinnitus handicap inventory self-administered questionnaire, were performed before the therapy and at 3, 6, and 12 months during the therapy. Data showed a highly significant improvement in both tinnitus treatments starting from the first three months and up to one year of therapy, with a progressive and statistically significant decrease in the disability every three months. TRT was equally effective with sound generator or open ear hearing aids: they gave basically identical, statistically indistinguishable results.
Source and listener directivity for interactive wave-based sound propagation.
Mehra, Ravish; Antani, Lakulish; Kim, Sujeong; Manocha, Dinesh
2014-04-01
We present an approach to model dynamic, data-driven source and listener directivity for interactive wave-based sound propagation in virtual environments and computer games. Our directional source representation is expressed as a linear combination of elementary spherical harmonic (SH) sources. In the preprocessing stage, we precompute and encode the propagated sound fields due to each SH source. At runtime, we perform the SH decomposition of the varying source directivity interactively and compute the total sound field at the listener position as a weighted sum of precomputed SH sound fields. We propose a novel plane-wave decomposition approach based on higher-order derivatives of the sound field that enables dynamic HRTF-based listener directivity at runtime. We provide a generic framework to incorporate our source and listener directivity in any offline or online frequency-domain wave-based sound propagation algorithm. We have integrated our sound propagation system in Valve's Source game engine and use it to demonstrate realistic acoustic effects such as sound amplification, diffraction low-passing, scattering, localization, externalization, and spatial sound, generated by wave-based propagation of directional sources and listener in complex scenarios. We also present results from our preliminary user study.
Hearing at low and infrasonic frequencies.
Møller, H; Pedersen, C S
2004-01-01
The human perception of sound at frequencies below 200 Hz is reviewed. Knowledge about our perception of this frequency range is important, since much of the sound we are exposed to in our everyday environment contains significant energy in this range. Sound at 20-200 Hz is called low-frequency sound, while for sound below 20 Hz the term infrasound is used. The hearing becomes gradually less sensitive for decreasing frequency, but despite the general understanding that infrasound is inaudible, humans can perceive infrasound, if the level is sufficiently high. The ear is the primary organ for sensing infrasound, but at levels somewhat above the hearing threshold it is possible to feel vibrations in various parts of the body. The threshold of hearing is standardized for frequencies down to 20 Hz, but there is a reasonably good agreement between investigations below this frequency. It is not only the sensitivity but also the perceived character of a sound that changes with decreasing frequency. Pure tones become gradually less continuous, the tonal sensation ceases around 20 Hz, and below 10 Hz it is possible to perceive the single cycles of the sound. A sensation of pressure at the eardrums also occurs. The dynamic range of the auditory system decreases with decreasing frequency. This compression can be seen in the equal-loudness-level contours, and it implies that a slight increase in level can change the perceived loudness from barely audible to loud. Combined with the natural spread in thresholds, it may have the effect that a sound, which is inaudible to some people, may be loud to others. Some investigations give evidence of persons with an extraordinary sensitivity in the low and infrasonic frequency range, but further research is needed in order to confirm and explain this phenomenon.
Johnson, Earl E; Light, Keri C
2015-09-01
To evaluate sound quality preferences of participants wearing hearing aids with different strengths of nonlinear frequency compression (NFC) processing versus no NFC processing. Two analysis methods, one without and one with a qualifier as to the magnitude of preferences, were compared for their percent agreement to differentiate a small difference in perceived sound quality as a result of applied NFC processing. A single-blind design was used with participants unaware of the presence or strength of NFC processing (independent variable). The National Acoustic Laboratories-Nonlinear 2 (NAL-NL2) prescription of amplification was chosen because audibility is intentionally not prescribed in the presence of larger sensorineural hearing loss thresholds. A lack of prescribed audibility, when present, was deemed an objective qualifier for NFC. NFC is known to improve the input bandwidth available to listeners when high-frequency audibility is not otherwise available and increasing strengths of NFC were examined. Experimental condition 3 (EC3) was stronger than the manufacturer default (EC2). More aggressive strengths (e.g., EC4 and EC5), however, were expected to include excessive distortion and even reduce the output bandwidth that had been prescribed as audible by NAL-NL2 (EC1). A total of 14 male Veterans with severe high-frequency sensorineural hearing loss. Participant sound quality preference ratings (dependent variable) without a qualifier as to the magnitude of preference were analyzed based on binomial probability theory, as is traditional with paired comparison data. The ratings with a qualifier as to the magnitude of preference were analyzed based on the nonparametric statistic of the Wilcoxon signed rank test. The binomial probability analysis method identified a sound quality preference as well as the nonparametric probability test method. As the strength of NFC increased, more participants preferred the EC with less NFC. Fourteen of 14 participants showed equal preference between EC1 and EC2 perhaps, in part, because EC2 showed no objective improvement in audibility for six of the 14 participants (42%). Thirteen of the 14 participants showed no preference between NAL-NL2 and EC3, but all participants had an objective improvement in audibility. With more NFC than EC3, more and more participants preferred the other EC with less NFC in the paired comparison. By referencing the recommended sensation levels of amplitude compression (e.g., NAL-NL2) in the ear canal of hearing aid wearers, the targeting of NFC parameters can likely be optimized with respect to improvements in effective audibility that may contribute to speech recognition without adversely impacting sound quality. After targeting of NFC parameters, providers can facilitate decisions about the use of NFC parameters (strengths of processing) via sound quality preference judgments using paired comparisons. American Academy of Audiology.
Semicircular Canal Pressure Changes During High-intensity Acoustic Stimulation.
Maxwell, Anne K; Banakis Hartl, Renee M; Greene, Nathaniel T; Benichoux, Victor; Mattingly, Jameson K; Cass, Stephen P; Tollin, Daniel J
2017-08-01
Acoustic stimulation generates measurable sound pressure levels in the semicircular canals. High-intensity acoustic stimuli can cause hearing loss and balance disruptions. To examine the propagation of acoustic stimuli to the vestibular end-organs, we simultaneously measured fluid pressure in the cochlea and semicircular canals during both air- and bone-conducted sound presentation. Five full-cephalic human cadaveric heads were prepared bilaterally with a mastoidectomy and extended facial recess. Vestibular pressures were measured within the superior, lateral, and posterior semicircular canals, and referenced to intracochlear pressure within the scala vestibuli with fiber-optic pressure probes. Pressures were measured concurrently with laser Doppler vibrometry measurements of stapes velocity during stimulation with both air- and bone-conduction. Stimuli were pure tones between 100 Hz and 14 kHz presented with custom closed-field loudspeakers for air-conducted sounds and via commercially available bone-anchored device for bone-conducted sounds. Pressures recorded in the superior, lateral, and posterior semicircular canals in response to sound stimulation were equal to or greater in magnitude than those recorded in the scala vestibuli (up to 20 dB higher). The pressure magnitudes varied across canals in a frequency-dependent manner. High sound pressure levels were recorded in the semicircular canals with sound stimulation, suggesting that similar acoustical energy is transmitted to the semicircular canals and the cochlea. Since these intralabyrinthine pressures exceed intracochlear pressure levels, our results suggest that the vestibular end-organs may also be at risk for injury during exposure to high-intensity acoustic stimuli known to cause trauma in the auditory system.
Hamerschmidt, Rogério; Schuch, Luiz Henrique; Rezende, Rodrigo Kopp; Wiemes, Gislaine Richter Minhoto; Oliveira, Adriana Kosma Pires de; Mocellin, Marcos
2012-01-01
There are two techniques for cochlear implant (CI) electrode placement: cochleostomy and the round window (RW) approach. This study aims to compare neural response telemetry (NRT) results immediately after surgery to check for possible differences on auditory nerve stimulation between these two techniques. This is a prospective cross-sectional study. Twenty-three patients were enrolled. Six patients underwent surgery by cochleostomy and 17 had it through the RW approach. Mean charge units (MCU) for high frequency sounds: patients submitted to the RW approach had a mean value of 190.4 (± 29.2) while cochleostomy patients averaged 187.8 (± 32.7); p = 0.71. MCU for mid frequency sounds: patients submitted to the RW approach had a mean value of 192.5 (± 22) while cochleostomy patients averaged 178.5 (± 18.5); p = 0.23. MCU for low frequency sounds: patients submitted to the RW approach had a mean value of 183.3 (± 25) while cochleostomy patients averaged 163.8 (± 19.3); p = 0.19. This study showed no differences in the action potential of the distal portion of the auditory nerve in patients with multichannel cochlear implants submitted to surgery by cochleostomy or through the RW approach, using the implant itself to generate stimuli and record responses. Both techniques equally stimulate the cochlear nerve. Therefore, the choice of approach can be made based on the surgeon's own preference and experience.
U. S. Geological Survey programs in Michigan
,
1996-01-01
For more than 100 years, the United States Geological Survey (USGS) has provided earth-science information on which managers, scientists, and other interested citizens base decisions regarding Michigan’s natural resources and natural hazards. The non-regulatory and scientific nature of the USGS work ensures that our products are technically sound, unbiased, and equally accessible and available to all interested parties. The various programs of the USGS in Michigan reflect a response to the citizens of Michigan and their need for geologic, topographic, biologic, and hydrologic information. Much of the work of the USGS in Michigan is part of cooperative programs in which the diversity of interests among local, regional, State, Tribal, and Federal agencies is accommodated through joint planning and funding.
Supporting Blind Students in STEM Education in Austria.
Petz, Andrea; Miesenberger, Klaus
2015-01-01
Equal access to education will foster a knowledge society for all. In particular for the ICT based information society a benchmark has been set to raise the numbers of graduates in science, technology, engineering and mathematics (STEM) study courses by 15% (748.000) per year, asking for increased efforts in Europe (http://ec.europa.eu/education/policy/strategic-framework/index_en.htm). This holds even more true for people with disabilities who a) participate in and graduate from STEM at a much lower number and b) face a much higher unemployment rate, in particular in STEM related fields. This asks for sound and well-founded education - first and foremost in math - for people with disability and here especially for blind people.
A note on sound radiation from distributed sources
NASA Technical Reports Server (NTRS)
Levine, H.
1979-01-01
The power output from a normally vibrating strip radiator is expressed in alternative general forms, one of these being chosen to refine and correct some particular estimates given by Heckl for different numerical ratios of strip width to wave length. An exact and explicit calculation is effected for sinusoidal velocity profiles when the strip width equals an integer number of half wave lengths.
Decision making for wildfires: A guide for applying a risk management process at the incident level
Mary A. Taber; Lisa M. Elenz; Paul G. Langowski
2013-01-01
This publication focuses on the thought processes and considerations surrounding a risk management process for decision making on wildfires. The publication introduces a six element risk management cycle designed to encourage sound risk-informed decision making in accordance with Federal wildland fire policy, although the process is equally applicable to non-Federal...
Subwavelength Focalization of Acoustic Waves Using Time Reversal. Yes We Can!
ERIC Educational Resources Information Center
El Abed, Mohamed
2014-01-01
By superimposing two sound waves of the same wavelength, propagating in the opposite direction, we can create an intensity pattern having a characteristic scale equal to half a wavelength: it is the diffraction limit. Recently a group from the Institut Laue-Langevin in Paris has shown that it is possible to go beyond this limit by focusing sound…
Pitch perception deficits in nonverbal learning disability.
Fernández-Prieto, I; Caprile, C; Tinoco-González, D; Ristol-Orriols, B; López-Sala, A; Póo-Argüelles, P; Pons, F; Navarra, J
2016-12-01
The nonverbal learning disability (NLD) is a neurological dysfunction that affects cognitive functions predominantly related to the right hemisphere such as spatial and abstract reasoning. Previous evidence in healthy adults suggests that acoustic pitch (i.e., the relative difference in frequency between sounds) is, under certain conditions, encoded in specific areas of the right hemisphere that also encode the spatial elevation of external objects (e.g., high vs. low position). Taking this evidence into account, we explored the perception of pitch in preadolescents and adolescents with NLD and in a group of healthy participants matched by age, gender, musical knowledge and handedness. Participants performed four speeded tests: a stimulus detection test and three perceptual categorization tests based on colour, spatial position and pitch. Results revealed that both groups were equally fast at detecting visual targets and categorizing visual stimuli according to their colour. In contrast, the NLD group showed slower responses than the control group when categorizing space (direction of a visual object) and pitch (direction of a change in sound frequency). This pattern of results suggests the presence of a subtle deficit at judging pitch in NLD along with the traditionally-described difficulties in spatial processing. Copyright © 2016. Published by Elsevier Ltd.
Improvement of impact noise in a passenger car utilizing sound metric based on wavelet transform
NASA Astrophysics Data System (ADS)
Lee, Sang-Kwon; Kim, Ho-Wuk; Na, Eun-Woo
2010-08-01
A new sound metric for impact sound is developed based on the continuous wavelet transform (CWT), a useful tool for the analysis of non-stationary signals such as impact noise. Together with new metric, two other conventional sound metrics related to sound modulation and fluctuation are also considered. In all, three sound metrics are employed to develop impact sound quality indexes for several specific impact courses on the road. Impact sounds are evaluated subjectively by 25 jurors. The indexes are verified by comparing the correlation between the index output and results of a subjective evaluation based on a jury test. These indexes are successfully applied to an objective evaluation for improvement of the impact sound quality for cases where some parts of the suspension system of the test car are modified.
Propagation of sound waves through a spatially homogeneous but smoothly time-dependent medium
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hayrapetyan, A.G., E-mail: armen@physi.uni-heidelberg.de; Max-Planck-Institut für Kernphysik, Saupfercheckweg 1, D-69117 Heidelberg; Grigoryan, K.K.
2013-06-15
The propagation of sound through a spatially homogeneous but non-stationary medium is investigated within the framework of fluid dynamics. For a non-vortical fluid, especially, a generalized wave equation is derived for the (scalar) potential of the fluid velocity distribution in dependence of the equilibrium mass density of the fluid and the sound wave velocity. A solution of this equation for a finite transition period τ is determined in terms of the hypergeometric function for a phenomenologically realistic, sigmoidal change of the mass density and sound wave velocity. Using this solution, it is shown that the energy flux of the soundmore » wave is not conserved but increases always for the propagation through a non-stationary medium, independent of whether the equilibrium mass density is increased or decreased. It is found, moreover, that this amplification of the transmitted wave arises from an energy exchange with the medium and that its flux is equal to the (total) flux of the incident and the reflected wave. An interpretation of the reflected wave as a propagation of sound backward in time is given in close analogy to Feynman and Stueckelberg for the propagation of anti-particles. The reflection and transmission coefficients of sound propagating through a non-stationary medium is analyzed in more detail for hypersonic waves with transition periods τ between 15 and 200 ps as well as the transformation of infrasound waves in non-stationary oceans. -- Highlights: •Analytically exact study of sound propagation through a non-stationary medium. •Energy exchange between the non-stationary medium and the sound wave. •Transformation of hypersonic and ultrasound frequencies in non-stationary media. •Propagation of sound backward in time in close analogy to anti-particles. •Prediction of tsunamis both in spatially and temporally inhomogeneous oceans.« less
The topography of frequency and time representation in primate auditory cortices
Baumann, Simon; Joly, Olivier; Rees, Adrian; Petkov, Christopher I; Sun, Li; Thiele, Alexander; Griffiths, Timothy D
2015-01-01
Natural sounds can be characterised by their spectral content and temporal modulation, but how the brain is organized to analyse these two critical sound dimensions remains uncertain. Using functional magnetic resonance imaging, we demonstrate a topographical representation of amplitude modulation rate in the auditory cortex of awake macaques. The representation of this temporal dimension is organized in approximately concentric bands of equal rates across the superior temporal plane in both hemispheres, progressing from high rates in the posterior core to low rates in the anterior core and lateral belt cortex. In A1 the resulting gradient of modulation rate runs approximately perpendicular to the axis of the tonotopic gradient, suggesting an orthogonal organisation of spectral and temporal sound dimensions. In auditory belt areas this relationship is more complex. The data suggest a continuous representation of modulation rate across several physiological areas, in contradistinction to a separate representation of frequency within each area. DOI: http://dx.doi.org/10.7554/eLife.03256.001 PMID:25590651
Deriving a dosage-response relationship for community response to high-energy impulsive noise
NASA Technical Reports Server (NTRS)
Fidell, Sanford; Pearsons, Karl S.
1994-01-01
The inability to systematically predict community response to exposure to sonic booms (and other high energy impulsive sounds) is a major impediment to credible analyses of the environmental effects of supersonic flight operations. Efforts to assess community response to high energy impulsive sounds are limited in at least two important ways. First, a paucity of appropriate empirical data makes it difficult to infer a dosage-response relationship by means similar to those used in the case of general transportation noise. Second, it is unclear how well the 'equal energy hypothesis' (the notion that duration, number, and level of individual events are directly interchangeable determinants of annoyance) applies to some forms of impulsive noise exposure. Some of the issues currently under consideration by a CHABA working group addressing these problems are discussed. These include means for applying information gained in controlled exposure studies about different rates of growth of annoyance with impulsive and non-impulsive sound exposure levels, and strategies for developing a dosage-response relationship in a data-poor area.
Leak locating microphone, method and system for locating fluid leaks in pipes
Kupperman, David S.; Spevak, Lev
1994-01-01
A leak detecting microphone inserted directly into fluid within a pipe includes a housing having a first end being inserted within the pipe and a second opposed end extending outside the pipe. A diaphragm is mounted within the first housing end and an acoustic transducer is coupled to the diaphragm for converting acoustical signals to electrical signals. A plurality of apertures are provided in the housing first end, the apertures located both above and below the diaphragm, whereby to equalize fluid pressure on either side of the diaphragm. A leak locating system and method are provided for locating fluid leaks within a pipe. A first microphone is installed within fluid in the pipe at a first selected location and sound is detected at the first location. A second microphone is installed within fluid in the pipe at a second selected location and sound is detected at the second location. A cross-correlation is identified between the detected sound at the first and second locations for identifying a leak location.
Musical rhythm and reading development: does beat processing matter?
Ozernov-Palchik, Ola; Patel, Aniruddh D
2018-05-20
There is mounting evidence for links between musical rhythm processing and reading-related cognitive skills, such as phonological awareness. This may be because music and speech are rhythmic: both involve processing complex sound sequences with systematic patterns of timing, accent, and grouping. Yet, there is a salient difference between musical and speech rhythm: musical rhythm is often beat-based (based on an underlying grid of equal time intervals), while speech rhythm is not. Thus, the role of beat-based processing in the reading-rhythm relationship is not clear. Is there is a distinct relation between beat-based processing mechanisms and reading-related language skills, or is the rhythm-reading link entirely due to shared mechanisms for processing nonbeat-based aspects of temporal structure? We discuss recent evidence for a distinct link between beat-based processing and early reading abilities in young children, and suggest experimental designs that would allow one to further methodically investigate this relationship. We propose that beat-based processing taps into a listener's ability to use rich contextual regularities to form predictions, a skill important for reading development. © 2018 New York Academy of Sciences.
NASA Technical Reports Server (NTRS)
Grosveld, F.; Vanaken, J.
1978-01-01
Sound pressure levels in the test facility were studied that are caused by varying: (1) microphone positions; (2) equalizer setting; and (3) panel clamping forces. Measurements were done by using a Beranek tube or this Beranek tube in combinations with an extension tube and a special test section. In all configurations tests were executed with and without a test panel installed. The influence of the speaker back panel and the back panel of the Beranek tube on the sound pressure levels inside the test tube were also investigated. It is shown that the definition of noise reduction is more useful in relation to this test facility than transmission loss.
An experimental study on target recognition using white canes.
Nunokawa, Kiyohiko; Ino, Shuichi
2010-01-01
To understand basic tactile perception using white canes, we compared tapping (two times) and pushing (two times) methods using the index finger and using a white cane, with and without accompanying auditory information. Participants were six visually impaired individuals who used a white cane to walk independently in their daily lives. For each of the tapping and pushing and sound or no sound conditions, participants gave magnitude estimates for the hardness of rubber panels. Results indicated that using a white cane produces sensitivity levels equal to using a finger when accompanied by auditory information, and suggested that when using a white cane to estimate the hardness of a target, it is most effective to have two different modalities of tactile and auditory information derived from tapping.
Kent, Shawn C.; Wanzek, Jeanne; Al Otaiba, Stephanie
2012-01-01
The purpose of this study was to examine the amount of time spent actively engaged in reading sounds, words, and connected text for students at-risk for reading difficulties in the first formal grade of reading instruction, kindergarten. Observational data of 109 kindergarten students at high-risk for later reading difficulties were collected during general education reading instruction across the school year. Findings revealed students read orally for just over 1 minute during their reading instruction with approximately equal time spent reading sounds, words, or connected text. Implications of these results for early reading instruction and intervention for students at-risk for reading difficulties or disabilities are presented. PMID:23087545
Sound absorption by subwavelength membrane structures: A geometric perspective
NASA Astrophysics Data System (ADS)
Yang, Min; Li, Yong; Meng, Chong; Fu, Caixing; Mei, Jun; Yang, Zhiyu; Sheng, Ping
2015-12-01
Decorated membranes comprising a thin layer of elastic film with small rigid platelets fixed on top have been found to be efficient absorbers of low-frequency sound. In this work we consider the problem of sound absorption from a perspective aimed at deriving upper bounds under different scenarios, i.e., whether the sound is incident from one side only or from both sides, and whether there is a reflecting surface on the back side of the membrane. By considering the negligible thickness of the membrane, usually on the order of a fraction of one millimeter, we derive a relation showing that the sum of the incoming sound waves' (complex) pressure amplitudes, averaged over the area of the membrane, must be equal to that of the outgoing waves. By using this relation, and without going to any details of the wave solutions, it is shown that the maximum absorption achievable from one-sided incidence is 50%, while the maximum absorption with a back-reflecting surface can reach 100%. The latter was attained by the hybridized resonances. All the results are shown to be in excellent agreement with the experiments. This generalized perspective, when used together with the Green function's formalism, can be useful in gaining insights into the constraints on what are achievable in scatterings and absorption by thin film structures and delineating them.
Neal, Christopher; Kennon-McGill, Stefanie; Freemyer, Andrea; Shum, Axel; Staecker, Hinrich; Durham, Dianne
2015-10-01
Exposure to intense sound can damage or kill cochlear hair cells (HC). This loss of input typically manifests as noise induced hearing loss, but it can also be involved in the initiation of other auditory disorders such as tinnitus or hyperacusis. In this study we quantify changes in HC number following exposure to one of four sound damage paradigms. We exposed adult, anesthetized Long-Evans rats to a unilateral 16 kHz pure tone that varied in intensity (114 dB or 118 dB) and duration (1, 2, or 4 h) and sacrificed animals 2-4 weeks later. We compared two different methods of tissue preparation, plastic embedding/sectioning and whole mount dissection, for quantifying hair cell loss as a function of frequency. We found that the two methods of tissue preparation produced largely comparable cochleograms, with whole mount dissections allowing a more rapid evaluation of hair cell number. Both inner and outer hair cell loss was observed throughout the length of the cochlea irrespective of sound damage paradigm. Inner HC loss was either equal to or greater than outer HC loss. Increasing the duration of sound exposures resulted in more severe HC loss, which included all HC lesions observed in an analogous shorter duration exposure. Copyright © 2015 Elsevier B.V. All rights reserved.
Redirection and Splitting of Sound Waves by a Periodic Chain of Thin Perforated Cylindrical Shells
NASA Astrophysics Data System (ADS)
Bozhko, Andrey; Sánchez-Dehesa, José; Cervera, Francisco; Krokhin, Arkadii
2017-06-01
The scattering of sound by finite and infinite chains of equally spaced perforated metallic cylindrical shells in an ideal (inviscid) and viscous fluid is theoretically studied using rigorous analytical and numerical approaches. Because of perforations, a chain of thin shells is practically transparent for sound within a wide range of frequencies. It is shown that strong scattering and redirection of sound by 90° may occur only for a discrete set of frequencies (Wood's anomalies) where the leaky eigenmodes are excited. The spectrum of eigenmodes consists of antisymmetric and symmetric branches with normal and anomalous dispersion, respectively. The antisymmetric eigenmode turns out to be a deaf mode, since it cannot be excited at normal incidence. However, at slightly oblique incidence, both modes can be resonantly excited at different but close frequencies. The symmetric mode, due to its anomalous dispersion, scatters sound in the "wrong" direction. This property may find an application for the splitting of the two resonant harmonics of the incoming signal into two beams propagating along the chain in the opposite directions. A chain of perforated cylinders may also be used as a passive antenna that detects the direction to the incoming signal by measuring the frequencies of the waves excited in the chain. Calculations are presented for aluminum shells in viscous air where the effects of anomalous scattering, redirection, and signal splitting are well manifested.
Development of an ICT-Based Air Column Resonance Learning Media
NASA Astrophysics Data System (ADS)
Purjiyanta, Eka; Handayani, Langlang; Marwoto, Putut
2016-08-01
Commonly, the sound source used in the air column resonance experiment is the tuning fork having disadvantage of unoptimal resonance results due to the sound produced which is getting weaker. In this study we made tones with varying frequency using the Audacity software which were, then, stored in a mobile phone as a source of sound. One advantage of this sound source is the stability of the resulting sound enabling it to produce the same powerful sound. The movement of water in a glass tube mounted on the tool resonance and the tone sound that comes out from the mobile phone were recorded by using a video camera. Sound resonances recorded were first, second, and third resonance, for each tone frequency mentioned. The resulting sound stays longer, so it can be used for the first, second, third and next resonance experiments. This study aimed to (1) explain how to create tones that can substitute tuning forks sound used in air column resonance experiments, (2) illustrate the sound wave that occurred in the first, second, and third resonance in the experiment, and (3) determine the speed of sound in the air. This study used an experimental method. It was concluded that; (1) substitute tones of a tuning fork sound can be made by using the Audacity software; (2) the form of sound waves that occured in the first, second, and third resonance in the air column resonance can be drawn based on the results of video recording of the air column resonance; and (3) based on the experiment result, the speed of sound in the air is 346.5 m/s, while based on the chart analysis with logger pro software, the speed of sound in the air is 343.9 ± 0.3171 m/s.
Multichannel feedforward control schemes with coupling compensation for active sound profiling
NASA Astrophysics Data System (ADS)
Mosquera-Sánchez, Jaime A.; Desmet, Wim; de Oliveira, Leopoldo P. R.
2017-05-01
Active sound profiling includes a number of control techniques that enables the equalization, rather than the mere reduction, of acoustic noise. Challenges may rise when trying to achieve distinct targeted sound profiles simultaneously at multiple locations, e.g., within a vehicle cabin. This paper introduces distributed multichannel control schemes for independently tailoring structural borne sound reaching a number of locations within a cavity. The proposed techniques address the cross interactions amongst feedforward active sound profiling units, which compensate for interferences of the primary sound at each location of interest by exchanging run-time data amongst the control units, while attaining the desired control targets. Computational complexity, convergence, and stability of the proposed multichannel schemes are examined in light of the physical system at which they are implemented. The tuning performance of the proposed algorithms is benchmarked with the centralized and pure-decentralized control schemes through computer simulations on a simplified numerical model, which has also been subjected to plant magnitude variations. Provided that the representation of the plant is accurate enough, the proposed multichannel control schemes have been shown as the only ones that properly deliver targeted active sound profiling tasks at each error sensor location. Experimental results in a 1:3-scaled vehicle mock-up further demonstrate that the proposed schemes are able to attain reductions of more than 60 dB upon periodic disturbances at a number of positions, while resolving cross-channel interferences. Moreover, when the sensor/actuator placement is found as defective at a given frequency, the inclusion of a regularization parameter in the cost function is seen to not hinder the proper operation of the proposed compensation schemes, at the time that it assures their stability, at the expense of losing control performance.
Nilsson, Mats E; Schenkman, Bo N
2016-02-01
Blind people use auditory information to locate sound sources and sound-reflecting objects (echolocation). Sound source localization benefits from the hearing system's ability to suppress distracting sound reflections, whereas echolocation would benefit from "unsuppressing" these reflections. To clarify how these potentially conflicting aspects of spatial hearing interact in blind versus sighted listeners, we measured discrimination thresholds for two binaural location cues: inter-aural level differences (ILDs) and inter-aural time differences (ITDs). The ILDs or ITDs were present in single clicks, in the leading component of click pairs, or in the lagging component of click pairs, exploiting processes related to both sound source localization and echolocation. We tested 23 blind (mean age = 54 y), 23 sighted-age-matched (mean age = 54 y), and 42 sighted-young (mean age = 26 y) listeners. The results suggested greater ILD sensitivity for blind than for sighted listeners. The blind group's superiority was particularly evident for ILD-lag-click discrimination, suggesting not only enhanced ILD sensitivity in general but also increased ability to unsuppress lagging clicks. This may be related to the blind person's experience of localizing reflected sounds, for which ILDs may be more efficient than ITDs. On the ITD-discrimination tasks, the blind listeners performed better than the sighted age-matched listeners, but not better than the sighted young listeners. ITD sensitivity declines with age, and the equal performance of the blind listeners compared to a group of substantially younger listeners is consistent with the notion that blind people's experience may offset age-related decline in ITD sensitivity. Copyright © 2015 The Authors. Published by Elsevier B.V. All rights reserved.
Development and modification of a Gaussian and non-Gaussian noise exposure system
NASA Astrophysics Data System (ADS)
Schlag, Adam W.
Millions of people across the world currently have noise induced hearing loss, and many are working in conditions with both continuous Gaussian and non-Gaussian noises that could affect their hearing. It was hypothesized that the energy of the noise was the cause of the hearing loss and did not depend on temporal pattern of a noise. This was referred to as the equal energy hypothesis. This hypothesis has been shown to have limitations though. This means that there is a difference in the types of noise a person receives to induce hearing loss and it is necessary to build a system that can easily mimic various conditions to conduct research. This study builds a system that can produce both non-Gaussian impulse/impact noises and continuous Gaussian noise. It was found that the peak sound pressure level of the system could reach well above the needed 120 dB level to represent acoustic trauma and could replicate well above the 85 dB A-weighted sound pressure level to produce conditions of gradual developing hearing loss. The system reached a maximum of 150 dB sound peak pressure level and a maximum of 133 dB A-weighted sound pressure level. Various parameters could easily be adjusted to control the sound, such as the high and low cutoff frequency to center the sound at 4 kHz. The system build can easily be adjusted to create numerous sound conditions and will hopefully be modified and improved in hopes of eventually being used for animal studies to lead to the creation of a method to treat or prevent noise induced hearing loss.
Pairwise comparisons and visual perceptions of equal area polygons.
Adamic, P; Babiy, V; Janicki, R; Kakiashvili, T; Koczkodaj, W W; Tadeusiewicz, R
2009-02-01
The number of studies related to visual perception has been plentiful in recent years. Participants rated the areas of five randomly generated shapes of equal area, using a reference unit area that was displayed together with the shapes. Respondents were 179 university students from Canada and Poland. The average error estimated by respondents using the unit square was 25.75%. The error was substantially decreased to 5.51% when the shapes were compared to one another in pairs. This gain of 20.24% for this two-dimensional experiment was substantially better than the 11.78% gain reported in the previous one-dimensional experiments. This is the first statistically sound two-dimensional experiment demonstrating that pairwise comparisons improve accuracy.
A simple computer-based measurement and analysis system of pulmonary auscultation sounds.
Polat, Hüseyin; Güler, Inan
2004-12-01
Listening to various lung sounds has proven to be an important diagnostic tool for detecting and monitoring certain types of lung diseases. In this study a computer-based system has been designed for easy measurement and analysis of lung sound using the software package DasyLAB. The designed system presents the following features: it is able to digitally record the lung sounds which are captured with an electronic stethoscope plugged to a sound card on a portable computer, display the lung sound waveform for auscultation sites, record the lung sound into the ASCII format, acoustically reproduce the lung sound, edit and print the sound waveforms, display its time-expanded waveform, compute the Fast Fourier Transform (FFT), and display the power spectrum and spectrogram.
Evaluating the Coda Phase Delay Method for Determining Temperature Ratios in Windy Environments
DOE Office of Scientific and Technical Information (OSTI.GOV)
Albert, Sarah; Bowman, Daniel; Rodgers, Arthur
2017-07-01
We evaluate the acoustic coda phase delay method for estimating changes in atmospheric phenomena in realistic environments. Previous studies verifying the method took place in an environment with negligible wind. The equation for effective sound speed, which the method is based upon, shows that the influence of wind is equal to the square of temperature. Under normal conditions, wind is significant and therefore cannot be ignored. Results from this study con rm the previous statement. The acoustic coda phase delay method breaks down in non-ideal environments, namely those where wind speed and direction varies across small distances. We suggest thatmore » future studies make use of gradiometry to better understand the effect of wind on the acoustic coda and subsequent phase delays.« less
Noise Reduction in Breath Sound Files Using Wavelet Transform Based Filter
NASA Astrophysics Data System (ADS)
Syahputra, M. F.; Situmeang, S. I. G.; Rahmat, R. F.; Budiarto, R.
2017-04-01
The development of science and technology in the field of healthcare increasingly provides convenience in diagnosing respiratory system problem. Recording the breath sounds is one example of these developments. Breath sounds are recorded using a digital stethoscope, and then stored in a file with sound format. This breath sounds will be analyzed by health practitioners to diagnose the symptoms of disease or illness. However, the breath sounds is not free from interference signals. Therefore, noise filter or signal interference reduction system is required so that breath sounds component which contains information signal can be clarified. In this study, we designed a filter called a wavelet transform based filter. The filter that is designed in this study is using Daubechies wavelet with four wavelet transform coefficients. Based on the testing of the ten types of breath sounds data, the data is obtained in the largest SNRdB bronchial for 74.3685 decibels.
Acoustic Calibration of the Exterior Effects Room at the NASA Langley Research Center
NASA Technical Reports Server (NTRS)
Faller, Kenneth J., II; Rizzi, Stephen A.; Klos, Jacob; Chapin, William L.; Surucu, Fahri; Aumann, Aric R.
2010-01-01
The Exterior Effects Room (EER) at the NASA Langley Research Center is a 39-seat auditorium built for psychoacoustic studies of aircraft community noise. The original reproduction system employed monaural playback and hence lacked sound localization capability. In an effort to more closely recreate field test conditions, a significant upgrade was undertaken to allow simulation of a three-dimensional audio and visual environment. The 3D audio system consists of 27 mid and high frequency satellite speakers and 4 subwoofers, driven by a real-time audio server running an implementation of Vector Base Amplitude Panning. The audio server is part of a larger simulation system, which controls the audio and visual presentation of recorded and synthesized aircraft flyovers. The focus of this work is on the calibration of the 3D audio system, including gains used in the amplitude panning algorithm, speaker equalization, and absolute gain control. Because the speakers are installed in an irregularly shaped room, the speaker equalization includes time delay and gain compensation due to different mounting distances from the focal point, filtering for color compensation due to different installations (half space, corner, baffled/unbaffled), and cross-over filtering.
Nordahl, Rolf; Turchet, Luca; Serafin, Stefania
2011-09-01
We propose a system that affords real-time sound synthesis of footsteps on different materials. The system is based on microphones, which detect real footstep sounds from subjects, from which the ground reaction force (GRF) is estimated. Such GRF is used to control a sound synthesis engine based on physical models. Two experiments were conducted. In the first experiment, the ability of subjects to recognize the surface they were exposed to was assessed. In the second experiment, the sound synthesis engine was enhanced with environmental sounds. Results show that, in some conditions, adding a soundscape significantly improves the recognition of the simulated environment.
Getting Home Safe and Sound: Occupational Safety and Health Administration at 38
Silverstein, Michael
2008-01-01
The Occupational Safety and Health Act of 1970 (OSHAct) declared that every worker is entitled to safe and healthful working conditions, and that employers are responsible for work being free from all recognized hazards. Thirty-eight years after these assurances, however, it is difficult to find anyone who believes the promise of the OSHAct has been met. The persistence of preventable, life-threatening hazards at work is a failure to keep a national promise. I review the history of the Occupational Safety and Health Administration and propose measures to better ensure that those who go to work every day return home safe and sound. These measures fall into 6 areas: leverage and accountability, safety and health systems, employee rights, equal protection, framing, and infrastructure. PMID:18235060
Getting home safe and sound: occupational safety and health administration at 38.
Silverstein, Michael
2008-03-01
The Occupational Safety and Health Act of 1970 (OSHAct) declared that every worker is entitled to safe and healthful working conditions, and that employers are responsible for work being free from all recognized hazards. Thirty-eight years after these assurances, however, it is difficult to find anyone who believes the promise of the OSHAct has been met. The persistence of preventable, life-threatening hazards at work is a failure to keep a national promise. I review the history of the Occupational Safety and Health Administration and propose measures to better ensure that those who go to work every day return home safe and sound. These measures fall into 6 areas: leverage and accountability, safety and health systems, employee rights, equal protection, framing, and infrastructure.
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-07-01
Measured and extrapolated data define the bioacoustic environments produced by a gasoline engine driven cabin leakage tester operating outdoors on a concrete apron at normal rated conditions. Near field data are presented for 37 locations at a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 - 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
Yost, William A; Zhong, Xuan; Najam, Anbar
2015-11-01
In four experiments listeners were rotated or were stationary. Sounds came from a stationary loudspeaker or rotated from loudspeaker to loudspeaker around an azimuth array. When either sounds or listeners rotate the auditory cues used for sound source localization change, but in the everyday world listeners perceive sound rotation only when sounds rotate not when listeners rotate. In the everyday world sound source locations are referenced to positions in the environment (a world-centric reference system). The auditory cues for sound source location indicate locations relative to the head (a head-centric reference system), not locations relative to the world. This paper deals with a general hypothesis that the world-centric location of sound sources requires the auditory system to have information about auditory cues used for sound source location and cues about head position. The use of visual and vestibular information in determining rotating head position in sound rotation perception was investigated. The experiments show that sound rotation perception when sources and listeners rotate was based on acoustic, visual, and, perhaps, vestibular information. The findings are consistent with the general hypotheses and suggest that sound source localization is not based just on acoustics. It is a multisystem process.
[Day-care mother--a successful model of child rearing?].
Schetelig, H
1977-02-17
The scientific reports published so far on the project "day care mothering" do not allow to draw scientific conclusions, whether this educational model can achieve nearly equal results as family education. The author critisises, that such a clame is made. It seems necessary, to pay more importance to the function of the mother for her children, to give the children a sound emotional basis for their future development.
The Auditory Hazard Assessment Algorithm for Humans (AHAAH): Hazard Evaluation of Intense Sounds
2011-07-01
6 3.2 Cochlear Susceptibility...pressures are higher (peak clipping during the higher pressure reduces the flow of energy to the inner ear) (Price, 2006). It has also been shown that...more or less energetic by 10 or 20 dB there is no assurance that the energy in the cochlear input will change by an equal amount. This explains why
How to generate a sound-localization map in fish
NASA Astrophysics Data System (ADS)
van Hemmen, J. Leo
2015-03-01
How sound localization is represented in the fish brain is a research field largely unbiased by theoretical analysis and computational modeling. Yet, there is experimental evidence that the axes of particle acceleration due to underwater sound are represented through a map in the midbrain of fish, e.g., in the torus semicircularis of the rainbow trout (Wubbels et al. 1997). How does such a map arise? Fish perceive pressure gradients by their three otolithic organs, each of which comprises a dense, calcareous, stone that is bathed in endolymph and attached to a sensory epithelium. In rainbow trout, the sensory epithelia of left and right utricle lie in the horizontal plane and consist of hair cells with equally distributed preferred orientations. We model the neuronal response of this system on the basis of Schuijf's vector detection hypothesis (Schuijf et al. 1975) and introduce a temporal spike code of sound direction, where optimality of hair cell orientation θj with respect to the acceleration direction θs is mapped onto spike phases via a von-Mises distribution. By learning to tune in to the earliest synchronized activity, nerve cells in the midbrain generate a map under the supervision of a locally excitatory, yet globally inhibitory visual teacher. Work done in collaboration with Daniel Begovic. Partially supported by BCCN - Munich.
Prediction of truly random future events using analysis of prestimulus electroencephalographic data
NASA Astrophysics Data System (ADS)
Baumgart, Stephen L.; Franklin, Michael S.; Jimbo, Hiroumi K.; Su, Sharon J.; Schooler, Jonathan
2017-05-01
Our hypothesis is that pre-stimulus physiological data can be used to predict truly random events tied to perceptual stimuli (e.g., lights and sounds). Our experiment presents light and sound stimuli to a passive human subject while recording electrocortical potentials using a 32-channel Electroencephalography (EEG) system. For every trial a quantum random number generator (qRNG) chooses from three possible selections with equal probability: a light stimulus, a sound stimulus, and no stimulus. Time epochs are defined preceding and post-ceding each stimulus for which mean average potentials were computed across all trials for the three possible stimulus types. Data from three regions of the brain are examined. In all three regions mean potential for light stimuli was generally enhanced relative to baseline during the period starting approximately 2 seconds before the stimulus. For sound stimuli, mean potential decreased relative to baseline during the period starting approximately 2 seconds before the stimulus. These changes from baseline may indicated the presence of evoked potentials arising from the stimulus. A P200 peak was observed in data recorded from frontal electrodes. The P200 is a well-known potential arising from the brain's processing of visual stimuli and its presence represents a replication of a known neurological phenomenon.
Nan, Yun; Liu, Li; Geiser, Eveline; Shu, Hua; Gong, Chen Chen; Dong, Qi; Gabrieli, John D E; Desimone, Robert
2018-06-25
Musical training confers advantages in speech-sound processing, which could play an important role in early childhood education. To understand the mechanisms of this effect, we used event-related potential and behavioral measures in a longitudinal design. Seventy-four Mandarin-speaking children aged 4-5 y old were pseudorandomly assigned to piano training, reading training, or a no-contact control group. Six months of piano training improved behavioral auditory word discrimination in general as well as word discrimination based on vowels compared with the controls. The reading group yielded similar trends. However, the piano group demonstrated unique advantages over the reading and control groups in consonant-based word discrimination and in enhanced positive mismatch responses (pMMRs) to lexical tone and musical pitch changes. The improved word discrimination based on consonants correlated with the enhancements in musical pitch pMMRs among the children in the piano group. In contrast, all three groups improved equally on general cognitive measures, including tests of IQ, working memory, and attention. The results suggest strengthened common sound processing across domains as an important mechanism underlying the benefits of musical training on language processing. In addition, although we failed to find far-transfer effects of musical training to general cognition, the near-transfer effects to speech perception establish the potential for musical training to help children improve their language skills. Piano training was not inferior to reading training on direct tests of language function, and it even seemed superior to reading training in enhancing consonant discrimination.
Miner, Nadine E.; Caudell, Thomas P.
2004-06-08
A sound synthesis method for modeling and synthesizing dynamic, parameterized sounds. The sound synthesis method yields perceptually convincing sounds and provides flexibility through model parameterization. By manipulating model parameters, a variety of related, but perceptually different sounds can be generated. The result is subtle changes in sounds, in addition to synthesis of a variety of sounds, all from a small set of models. The sound models can change dynamically according to changes in the simulation environment. The method is applicable to both stochastic (impulse-based) and non-stochastic (pitched) sounds.
Relational similarity-based model of data part 1: foundations and query systems
NASA Astrophysics Data System (ADS)
Belohlavek, Radim; Vychodil, Vilem
2017-10-01
We present a general rank-aware model of data which supports handling of similarity in relational databases. The model is based on the assumption that in many cases it is desirable to replace equalities on values in data tables by similarity relations expressing degrees to which the values are similar. In this context, we study various phenomena which emerge in the model, including similarity-based queries and similarity-based data dependencies. Central notion in our model is that of a ranked data table over domains with similarities which is our counterpart to the notion of relation on relation scheme from the classical relational model. Compared to other approaches which cover related problems, we do not propose a similarity-based or ranking module on top of the classical relational model. Instead, we generalize the very core of the model by replacing the classical, two-valued logic upon which the classical model is built by a more general logic involving a scale of truth degrees that, in addition to the classical truth degrees 0 and 1, contains intermediate truth degrees. While the classical truth degrees 0 and 1 represent nonequality and equality of values, and subsequently mismatch and match of queries, the intermediate truth degrees in the new model represent similarity of values and partial match of queries. Moreover, the truth functions of many-valued logical connectives in the new model serve to aggregate degrees of similarity. The presented approach is conceptually clean, logically sound, and retains most properties of the classical model while enabling us to employ new types of queries and data dependencies. Most importantly, similarity is not handled in an ad hoc way or by putting a "similarity module" atop the classical model in our approach. Rather, it is consistently viewed as a notion that generalizes and replaces equality in the very core of the relational model. We present fundamentals of the formal model and two equivalent query systems which are analogues of the classical relational algebra and domain relational calculus with range declarations. In the sequel to this paper, we deal with similarity-based dependencies.
Fuselage Structure Response to Boundary Layer, Tonal Sound, and Jet Noise
NASA Technical Reports Server (NTRS)
Maestrello, L.
2004-01-01
Experiments have been conducted to study the response of curved aluminum and graphite-epoxy fuselage structures to flow and sound loads from turbulent boundary layer, tonal sound, and jet noise. Both structures were the same size. The aluminum structure was reinforced with tear stoppers, while the graphite-epoxy structure was not. The graphite-epoxy structure weighed half as much as the aluminum structure. Spatiotemporal intermittence and chaotic behavior of the structural response was observed, as jet noise and tonal sound interacted with the turbulent boundary layer. The fundamental tone distributed energy to other components via wave interaction with the turbulent boundary layer. The added broadband sound from the jet, with or without a shock, influenced the responses over a wider range of frequencies. Instantaneous spatial correlation indicates small localized spatiotemporal regions of convected waves, while uncorrelated patterns dominate the larger portion of the space. By modifying the geometry of the tear stoppers between panels and frame, the transmitted and reflected waves of the aluminum panels were significantly reduced. The response level of the graphite-epoxy structure was higher, but the noise transmitted was nearly equal to that of the aluminum structure. The fundamental shock mode is between 80 deg and 150 deg and the first harmonic is between 20 deg and 80 deg for the underexpanded supersonic jet impinging on the turbulent boundary layer influencing the structural response. The response of the graphite-epoxy structure due to the fundamental mode of the shock impingement was stabilized by an externally fixed oscillator.
Tsai, Jang-Zern; Chang, Ming-Lang; Yang, Jiun-Yue; Kuo, Dar; Lin, Ching-Hsiung; Kuo, Cheng-Deng
2017-06-07
Though lung sounds auscultation is important for the diagnosis and monitoring of lung diseases, the spectral characteristics of lung sounds have not been fully understood. This study compared the spectral characteristics of lung sounds between the right and left lungs and between healthy male and female subjects using a dual-channel auscultation system. Forty-two subjects aged 18-22 years without smoking habits and any known pulmonary diseases participated in this study. The lung sounds were recorded from seven pairs of auscultation sites on the chest wall simultaneously. We found that in four out of seven auscultation pairs, the lung sounds from the left lung had a higher total power (P T ) than those from the right lung. The P T of male subjects was higher than that of female ones in most auscultation pairs. The ratio of inspiration power to expiration power (R I/E ) of lung sounds from the right lung was greater than that from the left lung at auscultation pairs on the anterior chest wall, while this phenomenon was reversed at auscultation pairs on the posterior chest wall in combined subjects, and similarly in both male and female subjects. Though the frequency corresponding to maximum power density of lung sounds (F MPD ) from the left and right lungs was not significantly different, the frequency that equally divided the power spectrum of lung sounds (F 50 ) from the left lung was significantly smaller than that from the right lung at auscultation site on the anterior and lateral chest walls, while it was significantly larger than that of from the right lung at auscultation site on the posterior chest walls. In conclusion, significant differences in the P T , F MPD , F 50 , and R I/E between the left and right lungs at some auscultation pairs were observed by using a dual-channel auscultation system in this study. Structural differences between the left and the right lungs, between the female and male subjects, and between anterior and posterior lungs might account for the observed differences in the spectral characteristics of lung sounds. The dual-channel auscultation system might be useful for future development of digital stethoscopes and power spectral analysis of lung sounds in patients with various kinds of cardiopulmonary diseases.
Tsai, Jang-Zern; Chang, Ming-Lang; Yang, Jiun-Yue; Kuo, Dar; Lin, Ching-Hsiung; Kuo, Cheng-Deng
2017-01-01
Though lung sounds auscultation is important for the diagnosis and monitoring of lung diseases, the spectral characteristics of lung sounds have not been fully understood. This study compared the spectral characteristics of lung sounds between the right and left lungs and between healthy male and female subjects using a dual-channel auscultation system. Forty-two subjects aged 18–22 years without smoking habits and any known pulmonary diseases participated in this study. The lung sounds were recorded from seven pairs of auscultation sites on the chest wall simultaneously. We found that in four out of seven auscultation pairs, the lung sounds from the left lung had a higher total power (PT) than those from the right lung. The PT of male subjects was higher than that of female ones in most auscultation pairs. The ratio of inspiration power to expiration power (RI/E) of lung sounds from the right lung was greater than that from the left lung at auscultation pairs on the anterior chest wall, while this phenomenon was reversed at auscultation pairs on the posterior chest wall in combined subjects, and similarly in both male and female subjects. Though the frequency corresponding to maximum power density of lung sounds (FMPD) from the left and right lungs was not significantly different, the frequency that equally divided the power spectrum of lung sounds (F50) from the left lung was significantly smaller than that from the right lung at auscultation site on the anterior and lateral chest walls, while it was significantly larger than that of from the right lung at auscultation site on the posterior chest walls. In conclusion, significant differences in the PT, FMPD, F50, and RI/E between the left and right lungs at some auscultation pairs were observed by using a dual-channel auscultation system in this study. Structural differences between the left and the right lungs, between the female and male subjects, and between anterior and posterior lungs might account for the observed differences in the spectral characteristics of lung sounds. The dual-channel auscultation system might be useful for future development of digital stethoscopes and power spectral analysis of lung sounds in patients with various kinds of cardiopulmonary diseases. PMID:28590447
Tan, Zhixiang; Zhang, Yi; Zeng, Deping; Wang, Hua
2015-04-01
We proposed a research of a heart sound envelope extraction system in this paper. The system was implemented on LabVIEW based on the Hilbert-Huang transform (HHT). We firstly used the sound card to collect the heart sound, and then implemented the complete system program of signal acquisition, pretreatment and envelope extraction on LabVIEW based on the theory of HHT. Finally, we used a case to prove that the system could collect heart sound, preprocess and extract the envelope easily. The system was better to retain and show the characteristics of heart sound envelope, and its program and methods were important to other researches, such as those on the vibration and voice, etc.
Laboratory study of effects of sonic boom shaping on subjective loudness and acceptability
NASA Technical Reports Server (NTRS)
Leatherwood, Jack D.; Sullivan, Brenda M.
1992-01-01
A laboratory study was conducted to determine the effects of sonic boom signature shaping on subjective loudness and acceptability. The study utilized the sonic boom simulator at the Langley Research Center. A wide range of symmetrical, front-shock-minimized signature shapes were investigated together with a limited number of asymmetrical signatures. Subjective loudness judgments were obtained from 60 test subjects by using an 11-point numerical category scale. Acceptability judgments were obtained using the method of constant stimuli. Results were used to assess the relative predictive ability of several noise metrics, determine the loudness benefits of detailed boom shaping, and derive laboratory sonic boom acceptability criteria. These results indicated that the A-weighted sound exposure level, the Stevens Mark 7 Perceived Level, and the Zwicker Loudness Level metrics all performed well. Significant reductions in loudness were obtained by increasing front-shock rise time and/or decreasing front-shock overpressure of the front-shock minimized signatures. In addition, the asymmetrical signatures were rated to be slightly quieter than the symmetrical front-shock-minimized signatures of equal A-weighted sound exposure level. However, this result was based on a limited number of asymmetric signatures. The comparison of laboratory acceptability results with acceptability data obtained in more realistic situations also indicated good agreement.
Park, Munhum
2013-07-01
In contrast to common expectations, the noise levels measured in hospital wards are known to be high with little day-night variation, potentially having negative effects on the patient outcomes and the work performance of the staff members, and considerable research attention has been drawn to such adverse acoustic conditions in healthcare environments. Recently, Young et al. [J. Acoust. Soc. Am. 132(5), 3234-3239 (2012)] proposed to use an artificial neural network (ANN) to predict the hourly energy-equivalent sound pressure level (Leq, 1h), particularly targeting neonatal intensive care units. Despite the timeliness of the study and the potential benefits of an "acoustic forecasting" model, the proposed scheme appears to be underdeveloped in a few important aspects, which this letter attempts to address. In particular, the prediction of a simpler time-series smoothing technique was equally or more accurate compared to that of the ANN. In addition, the percentage error used to indicate the prediction accuracy was not only perceptually irrelevant but also misleading given the narrow distribution of test data. Furthermore, this letter raises the more general question whether the sound pressure level may meaningfully be modeled solely based on the past time-series.
Huang, Lixi
2008-11-01
A spectral method of Chebyshev collocation with domain decomposition is introduced for linear interaction between sound and structure in a duct lined with flexible walls backed by cavities with or without a porous material. The spectral convergence is validated by a one-dimensional problem with a closed-form analytical solution, and is then extended to the two-dimensional configuration and compared favorably against a previous method based on the Fourier-Galerkin procedure and a finite element modeling. The nonlocal, exact Dirichlet-to-Neumann boundary condition is embedded in the domain decomposition scheme without imposing extra computational burden. The scheme is applied to the problem of high-frequency sound absorption by duct lining, which is normally ineffective when the wavelength is comparable with or shorter than the duct height. When a tensioned membrane covers the lining, however, it scatters the incident plane wave into higher-order modes, which then penetrate the duct lining more easily and get dissipated. For the frequency range of f=0.3-3 studied here, f=0.5 being the first cut-on frequency of the central duct, the membrane cover is found to offer an additional 0.9 dB attenuation per unit axial distance equal to half of the duct height.
Marquet, P
1995-01-01
After having invented the stethoscope, Laennec published his treatise on auscultation in 1819, describing the acoustic events generated by ventilation and linking them with anatomopathological findings. The weak points of his semiology lay in its subjective and interpretative character, expressed by an imprecise and picturesque nomenclature. Technical studies of breath sounds began in the middle of the twentieth century, and this enabled the American Thoracic Society to elaborate a new classification of adventitious noises based on a few physical characteristics. This terminology replaced that of Laennec or his translators (except in France). The waveforms of the different normal and adventitious noises have been well described. However, only the study of the time evolution of their tone (frequency-amplitude-time relationship) will enable a complete analysis of these phenomena. This approach has been undertaken by a few teams but much remains to be done, in particular in relation to discontinuous noises (crackles). Technology development raises hope for the design, in near future, of automatic processes for respiratory noise detection and classification. Systematic research into the production mechanisms and sites of these noises has progressed equally. It should, in time, reinforce their semiological value and give to auscultation, either instrumental or using the stethoscope or instrumentally, an increased diagnostic power and the status of respiratory function test.
Velocity perception for sounds moving in frequency space.
Henry, Molly J; McAuley, J Devin
2011-01-01
In three experiments, we considered the relative contribution of frequency change (Δf) and time change (Δt) to perceived velocity (Δf/Δt) for sounds that moved either continuously in frequency space (Experiment 1) or in discrete steps (Experiments 2 and 3). In all the experiments, participants estimated "how quickly stimuli changed in pitch" on a scale ranging from 0 (not changing at all) to 100 (changing very quickly). Objective frequency velocity was specified in terms of semitones per second (ST/s), with ascending and descending stimuli presented on each trial at one of seven velocities (2, 4, 6, 8, 10, 12, and 14 ST/s). Separate contributions of frequency change (Δf) and time change (Δt) to perceived velocity were assessed by holding total Δt constant and varying Δf or vice versa. For tone glides that moved continuously in frequency space, both Δf and Δt cues contributed approximately equally to perceived velocity. For tone sequences, in contrast, perceived velocity was based almost entirely on Δt, with surprisingly little contribution from Δf. Experiment 3 considered separate judgments about Δf and Δt in order to rule out the possibility that the results of Experiment 2 were due to the inability to judge frequency change in tone sequences.
Investigations of ionospheric sporadic Es layer using oblique sounding method
NASA Astrophysics Data System (ADS)
Minullin, R.
The characteristics of Es layer have been studied using oblique sounding at 28 radiolines at the frequencies of 34 -- 73 MHz at the transmission paths 400 -- 1600 km long during 30 years. Reflections from Es layer with a few hours duration were observed. The amplitude of the reflected signal reached 1000 μ V with the registration threshold 0,1 μ V. The borderlines between reflected and scattered signals were observed as sharp curves in 60 -- 100 s range on the distributions of duration of reflected signals for decameter waves. The duration of continuous Es reflections were decreased upon amplification of oblique sounding frequency. The distributions of duration of reflected signals for meter waves showed sharp curves in the range 200 -- 300 s, representing borderlines between signals reflected from meteoric traces and from Es layer. The filling coefficient for the oblique sounding as well as the Es layer emersion probability for the vertical sounding were shown to undergo daily, seasonal and periodic variations. The daily variations of the filling coefficient of Es signals showed clear-cut maximums at 10 -- 12 and 18 -- 20 hours and minimum at 4 -- 6 hours at all paths in summer time and the maximum at 12 -- 14 hours in winter time. The values of the filling coefficient for Es layer declined with the increase of oblique sounding frequency. The minimal values of the filling coefficient were observed in winter and early spring, while the maximal values were observed from May to August. Provided that the averaged filling coefficient is equal to one in summer, it reaches the level 0,25 in equinox and does not exceed the level 0,12 in winter as evident by the of oblique sounding. The filling coefficient relation to the value of the voltage detection threshold was approximated by power-mode law. The filling coefficients for summer period showed exponential relation with equivalent sounding frequencies. The experimental evidence was generalized in an analytical model. Using this model the averaged Es layer filling coefficients for particular season of the year can be forecasted in case of given sounding frequency, path length, and voltage threshold.
Cabrera, Alvaro Fuentes; Hoffmann, Pablo Faundez
2010-01-01
This study is focused on the single-trial classification of auditory event-related potentials elicited by sound stimuli from different spatial directions. Five naϊve subjects were asked to localize a sound stimulus reproduced over one of 8 loudspeakers placed in a circular array, equally spaced by 45°. The subject was seating in the center of the circular array. Due to the complexity of an eight classes classification, our approach consisted on feeding our classifier with two classes, or spatial directions, at the time. The seven chosen pairs were 0°, which was the loudspeaker directly in front of the subject, with all the other seven directions. The discrete wavelet transform was used to extract features in the time-frequency domain and a support vector machine performed the classification procedure. The average accuracy over all subjects and all pair of spatial directions was 76.5%, σ = 3.6. The results of this study provide evidence that the direction of a sound is encoded in single-trial auditory event-related potentials.
NASA Technical Reports Server (NTRS)
Harrington, W. W.
1973-01-01
The reduction is discussed of the discrete tones generated by jet engines which is essential for jet aircraft to meet present and proposed noise standards. The discrete tones generated by the blades and vanes propagate in the inlet and exhaust duct in the form of spiraling acoustic waves, or spinning modes. The reduction of these spinning modes by the cancellation effect of the combination of two acoustic fields was investigated. The spinning mode synthesizer provided the means for effective study of this noise reduction scheme. Two sets of electrical-acoustical transducers located in an equally-spaced circular array simultaneously generate a specified spinning mode and the cancelling mode. Analysis of the wave equation for the synthesizer established the optimum cancelling array acoustic parameters for maximum sound pressure level reduction. The parameter dependence of the frequency ranges of propagation of single, specified circumferential modes generated by a single array, and of effective cancellation of the modes generated by two arrays, was determined. Substantial sound pressure level reduction was obtained for modes within these limits.
Assessing Auditory Discrimination Skill of Malay Children Using Computer-based Method.
Ting, H; Yunus, J; Mohd Nordin, M Z
2005-01-01
The purpose of this paper is to investigate the auditory discrimination skill of Malay children using computer-based method. Currently, most of the auditory discrimination assessments are conducted manually by Speech-Language Pathologist. These conventional tests are actually general tests of sound discrimination, which do not reflect the client's specific speech sound errors. Thus, we propose computer-based Malay auditory discrimination test to automate the whole process of assessment as well as to customize the test according to the specific speech error sounds of the client. The ability in discriminating voiced and unvoiced Malay speech sounds was studied for the Malay children aged between 7 and 10 years old. The study showed no major difficulty for the children in discriminating the Malay speech sounds except differentiating /g/-/k/ sounds. Averagely the children of 7 years old failed to discriminate /g/-/k/ sounds.
A Flexible 360-Degree Thermal Sound Source Based on Laser Induced Graphene
Tao, Lu-Qi; Liu, Ying; Ju, Zhen-Yi; Tian, He; Xie, Qian-Yi; Yang, Yi; Ren, Tian-Ling
2016-01-01
A flexible sound source is essential in a whole flexible system. It’s hard to integrate a conventional sound source based on a piezoelectric part into a whole flexible system. Moreover, the sound pressure from the back side of a sound source is usually weaker than that from the front side. With the help of direct laser writing (DLW) technology, the fabrication of a flexible 360-degree thermal sound source becomes possible. A 650-nm low-power laser was used to reduce the graphene oxide (GO). The stripped laser induced graphene thermal sound source was then attached to the surface of a cylindrical bottle so that it could emit sound in a 360-degree direction. The sound pressure level and directivity of the sound source were tested, and the results were in good agreement with the theoretical results. Because of its 360-degree sound field, high flexibility, high efficiency, low cost, and good reliability, the 360-degree thermal acoustic sound source will be widely applied in consumer electronics, multi-media systems, and ultrasonic detection and imaging. PMID:28335239
NASA Astrophysics Data System (ADS)
Nishiura, Takanobu; Nakamura, Satoshi
2002-11-01
It is very important to capture distant-talking speech for a hands-free speech interface with high quality. A microphone array is an ideal candidate for this purpose. However, this approach requires localizing the target talker. Conventional talker localization algorithms in multiple sound source environments not only have difficulty localizing the multiple sound sources accurately, but also have difficulty localizing the target talker among known multiple sound source positions. To cope with these problems, we propose a new talker localization algorithm consisting of two algorithms. One is DOA (direction of arrival) estimation algorithm for multiple sound source localization based on CSP (cross-power spectrum phase) coefficient addition method. The other is statistical sound source identification algorithm based on GMM (Gaussian mixture model) for localizing the target talker position among localized multiple sound sources. In this paper, we particularly focus on the talker localization performance based on the combination of these two algorithms with a microphone array. We conducted evaluation experiments in real noisy reverberant environments. As a result, we confirmed that multiple sound signals can be identified accurately between ''speech'' or ''non-speech'' by the proposed algorithm. [Work supported by ATR, and MEXT of Japan.
Pascotini, Fernanda dos Santos; Ribeiro, Vanessa Veis; Christmann, Mara Keli; Tomasi, Lidia Lis; Dellazzana, Amanda Alves; Haeffner, Leris Salete Bonfanti; Cielo, Carla Aparecida
2016-01-01
Relate respiratory muscle strength (RMS), sound pressure (SP) level, and vocal acoustic parameters to the abdominal circumference (AC) and nutritional status of children. This is a cross-sectional study. Eighty-two school children aged between 8 and 10 years, grouped by nutritional states (eutrophic, overweight, or obese) and AC percentile (≤25, 25-75, and ≥75), were included in the study. Evaluations of maximal inspiratory pressure (IPmax) and maximal expiratory pressure (EPmax) were conducted using the manometer and SP and acoustic parameters through the Multi-Dimensional Voice Program Advanced (KayPENTAX, Montvale, New Jersey). There were significant differences (P < 0.05) in the EPmax of children with AC between the 25th and 75th percentiles (72.4) and those less than or equal to the 25th percentile (61.9) and in the SP of those greater than or equal to the 75th percentile (73.4) and less than or equal to the 25th percentile (66.6). The IPmax, EPmax, SP levels, and acoustic variables were not different in relation to the nutritional states of the children. There was a strong and positive correlation between the coefficient of amplitude perturbations (shimmer), the harmonics-to-noise ratio and the variation of the fundamental frequency, respectively, 0.79 and 0.71. RMS and acoustic voice characteristics in children do not appear to be influenced by nutritional states, and respiratory pressure does not interfere with acoustic voice characteristics. However, localized fat, represented by the AC, alters the EPmax and the SP, each of which increases as the AC increases. Copyright © 2016 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Deymier, P A; Swinteck, N; Runge, K; Deymier-Black, A; Hoying, J B
2015-01-01
We present a previously unrecognized effect of sound waves on gap-junction-based intercellular signaling such as in biological tissues composed of endothelial cells. We suggest that sound irradiation may, through temporal and spatial modulation of cell-to-cell conductance, create intercellular calcium waves with unidirectional signal propagation associated with nonconventional topologies. Nonreciprocity in calcium wave propagation induced by sound wave irradiation is demonstrated in the case of a linear and a nonlinear reaction-diffusion model. This demonstration should be applicable to other types of gap-junction-based intercellular signals, and it is thought that it should be of help in interpreting a broad range of biological phenomena associated with the beneficial therapeutic effects of sound irradiation and possibly the harmful effects of sound waves on health.
Hasnain, Syed Farid-ul; Johansson, Eva; Mogren, Ingrid; Krantz, Gunilla
2012-04-28
The purpose of this qualitative study was to explore perceptions and expectations of young males and females, in Karachi, Pakistan, regarding their life prospects and gender roles, with resulting implications for health behaviour. The main theme emerging was "Young adults' prospects in life are hampered by psychosocial and gender equality constraints". Gender inequality and the low status of women in society were described as major obstacles to the overall development. Persistent withholding of information to the younger generation on sexual and reproductive health issues was perceived to increase exposure to health risks, particularly sexually transmitted infections (STIs). The present study reveals new discourses on equality among young adults, pointing towards an increasing, sound interaction between the sexes and aspirations for more gender equal relationships. Such views and awareness among the younger generation constitutes a strong force towards change of traditional norms, including reproductive health behaviour, and calls for policy change.
1976-01-28
source-receiver geometry dynamics. For a given time instant, each of the subroutines outputs time variables ( emission time, arrival time...transmission loss, depression/elevation and azimuthal arrival angles, received frequency and range variables (range at emission time, range at arrival time...with the wind equal 24.5 kts. In the double bottom bounce regions, the emission angles (at the virtual surface source) are moderately small (15
A Study of Underwater Sound Ray Tracing Methodology
1990-09-01
SRER ,LB,SV,SVU2,SVU,SU2,SU4,GI,G PIE = 3.14159265359D0 77 lEST = 0 MI = M C DISTRIBUTE THE K SOURCES EQUALLY AROUND THE CIRCLE COUNTER C CLOCKWISE...C(I) - TI-R(I TIMER(I = TIMCO) - T(5) HER(IW HC(I - Hl ZER(IW ZC(I - ZI SR = DSQRT(HI**2 + WA - Zl)**2) SRER = DSQRT(HC(I)**2 + (A2 - Zl)**2) - SR 50
The "fourth dimension" of gene transcription.
O'Malley, Bert W
2009-05-01
The three dimensions of space provide our relationship to position on the earth, but the fourth dimension of time has an equally profound influence on our lives. Everything from light and sound to weather and biology operate on the principle of measurable temporal periodicity. Consequently, a wide variety of time clocks affect all aspects of our existence. The annual (and biannual) cycles of activity, metabolism, and mating, the monthly physiological clocks of women and men, and the 24-h diurnal rhythms of humans are prime examples. Should it be surprising to us that the fourth dimension also impinges upon gene expression and that the genome itself is regulated by the fastest running of all biological clocks? Recent evidence substantiates the existence of such a ubiquitin-dependent transcriptional clock that is based upon the activation and destruction of transcriptional coactivators.
The “Fourth Dimension” of Gene Transcription
O'Malley, Bert W.
2009-01-01
The three dimensions of space provide our relationship to position on the earth, but the fourth dimension of time has an equally profound influence on our lives. Everything from light and sound to weather and biology operate on the principle of measurable temporal periodicity. Consequently, a wide variety of time clocks affect all aspects of our existence. The annual (and biannual) cycles of activity, metabolism, and mating, the monthly physiological clocks of women and men, and the 24-h diurnal rhythms of humans are prime examples. Should it be surprising to us that the fourth dimension also impinges upon gene expression and that the genome itself is regulated by the fastest running of all biological clocks? Recent evidence substantiates the existence of such a ubiquitin-dependent transcriptional clock that is based upon the activation and destruction of transcriptional coactivators. PMID:19221049
Effect of Process Variables on the Inertia Friction Welding of Superalloys LSHR and Mar-M247
NASA Astrophysics Data System (ADS)
Mahaffey, D. W.; Senkov, O. N.; Shivpuri, R.; Semiatin, S. L.
2016-08-01
The effect of inertia friction welding process parameters on microstructure evolution, weld plane quality, and the tensile behavior of welds between dissimilar nickel-base superalloys was established. For this purpose, the fine-grain, powder metallurgy alloy LSHR was joined to coarse-grain cast Mar-M247 using a fixed level of initial kinetic energy, but different combinations of the flywheel moment of inertia and initial rotation speed. It was found that welds made with the largest moment of inertia resulted in a sound bond with the best microstructure and room-temperature tensile strength equal to or greater than that of the parent materials. A relationship between the moment of inertia and weld process efficiency was established. The post-weld tensile behavior was interpreted in the context of observed microstructure gradients and weld-line defects.
Continuous robust sound event classification using time-frequency features and deep learning
Song, Yan; Xiao, Wei; Phan, Huy
2017-01-01
The automatic detection and recognition of sound events by computers is a requirement for a number of emerging sensing and human computer interaction technologies. Recent advances in this field have been achieved by machine learning classifiers working in conjunction with time-frequency feature representations. This combination has achieved excellent accuracy for classification of discrete sounds. The ability to recognise sounds under real-world noisy conditions, called robust sound event classification, is an especially challenging task that has attracted recent research attention. Another aspect of real-word conditions is the classification of continuous, occluded or overlapping sounds, rather than classification of short isolated sound recordings. This paper addresses the classification of noise-corrupted, occluded, overlapped, continuous sound recordings. It first proposes a standard evaluation task for such sounds based upon a common existing method for evaluating isolated sound classification. It then benchmarks several high performing isolated sound classifiers to operate with continuous sound data by incorporating an energy-based event detection front end. Results are reported for each tested system using the new task, to provide the first analysis of their performance for continuous sound event detection. In addition it proposes and evaluates a novel Bayesian-inspired front end for the segmentation and detection of continuous sound recordings prior to classification. PMID:28892478
Continuous robust sound event classification using time-frequency features and deep learning.
McLoughlin, Ian; Zhang, Haomin; Xie, Zhipeng; Song, Yan; Xiao, Wei; Phan, Huy
2017-01-01
The automatic detection and recognition of sound events by computers is a requirement for a number of emerging sensing and human computer interaction technologies. Recent advances in this field have been achieved by machine learning classifiers working in conjunction with time-frequency feature representations. This combination has achieved excellent accuracy for classification of discrete sounds. The ability to recognise sounds under real-world noisy conditions, called robust sound event classification, is an especially challenging task that has attracted recent research attention. Another aspect of real-word conditions is the classification of continuous, occluded or overlapping sounds, rather than classification of short isolated sound recordings. This paper addresses the classification of noise-corrupted, occluded, overlapped, continuous sound recordings. It first proposes a standard evaluation task for such sounds based upon a common existing method for evaluating isolated sound classification. It then benchmarks several high performing isolated sound classifiers to operate with continuous sound data by incorporating an energy-based event detection front end. Results are reported for each tested system using the new task, to provide the first analysis of their performance for continuous sound event detection. In addition it proposes and evaluates a novel Bayesian-inspired front end for the segmentation and detection of continuous sound recordings prior to classification.
Characteristic sounds facilitate visual search.
Iordanescu, Lucica; Guzman-Martinez, Emmanuel; Grabowecky, Marcia; Suzuki, Satoru
2008-06-01
In a natural environment, objects that we look for often make characteristic sounds. A hiding cat may meow, or the keys in the cluttered drawer may jingle when moved. Using a visual search paradigm, we demonstrated that characteristic sounds facilitated visual localization of objects, even when the sounds carried no location information. For example, finding a cat was faster when participants heard a meow sound. In contrast, sounds had no effect when participants searched for names rather than pictures of objects. For example, hearing "meow" did not facilitate localization of the word cat. These results suggest that characteristic sounds cross-modally enhance visual (rather than conceptual) processing of the corresponding objects. Our behavioral demonstration of object-based cross-modal enhancement complements the extensive literature on space-based cross-modal interactions. When looking for your keys next time, you might want to play jingling sounds.
Recurring patterns in the songs of humpback whales (Megaptera novaeangliae).
Green, Sean R; Mercado, Eduardo; Pack, Adam A; Herman, Louis M
2011-02-01
Humpback whales, unlike most mammalian species, learn new songs as adults. Populations of singers progressively and collectively change the sounds and patterns within their songs throughout their lives and across generations. In this study, humpback whale songs recorded in Hawaii from 1985 to 1995 were analyzed using self-organizing maps (SOMs) to classify the sounds within songs, and to identify sound patterns that were present across multiple years. These analyses supported the hypothesis that recurring, persistent patterns exist within whale songs, and that these patterns are defined at least in part by acoustic relationships between adjacent sounds within songs. Sound classification based on acoustic differences between adjacent sounds yielded patterns within songs that were more consistent from year to year than classifications based on the properties of single sounds. Maintenance of fixed ratios of acoustic modulation across sounds, despite large variations in individual sounds, suggests intrinsic constraints on how sounds change within songs. Such acoustically invariant cues may enable whales to recognize and assess variations in songs despite propagation-related distortion of individual sounds and yearly changes in songs. Copyright © 2011 Elsevier B.V. All rights reserved.
Research on fiber Bragg grating heart sound sensing and wavelength demodulation method
NASA Astrophysics Data System (ADS)
Zhang, Cheng; Miao, Chang-Yun; Gao, Hua; Gan, Jing-Meng; Li, Hong-Qiang
2010-11-01
Heart sound includes a lot of physiological and pathological information of heart and blood vessel. Heart sound detecting is an important method to gain the heart status, and has important significance to early diagnoses of cardiopathy. In order to improve sensitivity and reduce noise, a heart sound measurement method based on fiber Bragg grating was researched. By the vibration principle of plane round diaphragm, a heart sound sensor structure of fiber Bragg grating was designed and a heart sound sensing mathematical model was established. A formula of heart sound sensitivity was deduced and the theoretical sensitivity of the designed sensor is 957.11pm/KPa. Based on matched grating method, the experiment system was built, by which the excursion of reflected wavelength of the sensing grating was detected and the information of heart sound was obtained. Experiments show that the designed sensor can detect the heart sound and the reflected wavelength variety range is about 70pm. When the sampling frequency is 1 KHz, the extracted heart sound waveform by using the db4 wavelet has the same characteristics with a standard heart sound sensor.
The Sound of the Microwave Background
NASA Astrophysics Data System (ADS)
Whittle, M.
2004-05-01
One of the most impressive developments in modern cosmology has been the measurement and analysis of the tiny fluctuations seen in the cosmic microwave background (CMB) radiation. When discussing these fluctuations, cosmologists frequently refer to their acoustic nature -- sound waves moving through the hot gas appear as peaks and troughs when they cross the surface of last scattering. As is now well known, recent observations quantify the amplitudes of these waves over several octaves, revealing a fundamental tone with several harmonics, whose relative strengths and pitches reveal important cosmological parameters, including global curvature. Not surprisingly, these results have wonderful pedagogical value in educating and inspiring both students and the general public. To further enhance this educational experience, I have attempted what might seem rather obvious, namely converting the CMB power spectrum into an audible sound. By raising the pitch some 50 octaves so that the fundamental falls at 200 Hz (matching its harmonic ``l" value), we hear the resulting sound as a loud hissing roar. Matching the progress in observational results has been an equally impressive development of the theoretical treatment of CMB fluctuations. Using available computer simulations (e.g. CMBFAST) it is possible to recreate the subtley different sounds generated by different kinds of universe (e.g. different curvature or baryon content). Pushing further, one can generate the ``true" sound, characterized by P(k), rather than the ``observed" sound, characterized by C(l). From P(k), we learn that the fundamental and harmonics are offset, yielding a chord somewhere between a major and minor third. A sequence of models also allows one to follow the growth of sound during the first megayear: a descending scream, changing into a deepening roar, with subsequent growing hiss; matching the increase in wavelength caused by universal expansion, followed by the post recombination flow of gas into the small scale potential wells created by dark matter. This final sound, of course, sets the stage for all subsequent growth of cosmic structure, from stars (hiss), through galaxies (mid-range tones), to large scale structure (bass notes). Although popular presentations of CMB studies already make use of many visual and conceptual aids, introducing sound into the pedagogical mix can significantly enhance both the intellectual and the emotional impact of the subject on its audience.
Categorization of extremely brief auditory stimuli: domain-specific or domain-general processes?
Bigand, Emmanuel; Delbé, Charles; Gérard, Yannick; Tillmann, Barbara
2011-01-01
The present study investigated the minimum amount of auditory stimulation that allows differentiation of spoken voices, instrumental music, and environmental sounds. Three new findings were reported. 1) All stimuli were categorized above chance level with 50 ms-segments. 2) When a peak-level normalization was applied, music and voices started to be accurately categorized with 20 ms-segments. When the root-mean-square (RMS) energy of the stimuli was equalized, voice stimuli were better recognized than music and environmental sounds. 3) Further psychoacoustical analyses suggest that the categorization of extremely brief auditory stimuli depends on the variability of their spectral envelope in the used set. These last two findings challenge the interpretation of the voice superiority effect reported in previously published studies and propose a more parsimonious interpretation in terms of an emerging property of auditory categorization processes.
A case report of the transport of an infant with a tension pneumopericardium.
Mordue, Barbara C
2005-08-01
Neonatal pneumopericardium is a potentially fatal complication of positive-pressure ventilation and has become rare with the advent of surfactant replacement therapy. The clinical diagnosis, stabilization, treatment, and nursing care of an infant with pneumopericardium has not previously been discussed in the nursing literature. In this case report, delays in the recognition and definitive treatment of the pneumopericardium were encountered, resulting in the transport of an infant with a tension pneumopericardium and pneumoperitoneum. Root-cause analysis is used to identify contributing factors and examine system changes necessary to prevent the transport of another patient with a similar potentially life-threatening condition. Pneumopericardium should be suspected in any infant with an acute deterioration, especially in the presence of normal, equal breath sounds and muffled heart sounds, because prompt recognition and definitive treatment may be life-saving.
ERIC Educational Resources Information Center
Eshach, Haim
2014-01-01
This article describes the development and field test of the Sound Concept Inventory Instrument (SCII), designed to measure middle school students' concepts of sound. The instrument was designed based on known students' difficulties in understanding sound and the history of science related to sound and focuses on two main aspects of sound: sound…
Dr. Seuss's Sound Words: Playing with Phonics and Spelling.
ERIC Educational Resources Information Center
Gardner, Traci
Boom! Br-r-ring! Cluck! Moo!--exciting sounds are everywhere. Whether visiting online sites that play sounds or taking a "sound hike," ask your students to notice the sounds they hear, then write their own book, using sound words, based on Dr. Seuss's "Mr. Brown Can MOO! Can You?" During the three 45-minute sessions, grade K-2…
Germany's long-term-care insurance: putting a social insurance model into practice.
Geraedts, M; Heller, G V; Harrington, C A
2000-01-01
A growing population of elderly has intensified the demand for long-term care (LTC) services. In response to the mounting need, Germany put into effect a LTC Insurance Act in 1995 that introduced mandatory public or private LTC insurance for the entire population of 82 million. The program was based on the organizational principles that define the German social insurance system. Those individuals in the public system and their employers each pay contributions equal to 0.85 percent of each employee's gross wages or salary. Ten percent of the population with the highest incomes have chosen the option of purchasing private long term care insurance. Provisions were made for uniform eligibility criteria, benefits based on level of care needs, cost containment, and quality assurance. Over the first four years of its operation, the system has proved financially sound and has expanded access to organized LTC services. The German system thus may serve as an example for other countries that are planning to initiate social LTC insurance systems in other nations.
Incremental Query Rewriting with Resolution
NASA Astrophysics Data System (ADS)
Riazanov, Alexandre; Aragão, Marcelo A. T.
We address the problem of semantic querying of relational databases (RDB) modulo knowledge bases using very expressive knowledge representation formalisms, such as full first-order logic or its various fragments. We propose to use a resolution-based first-order logic (FOL) reasoner for computing schematic answers to deductive queries, with the subsequent translation of these schematic answers to SQL queries which are evaluated using a conventional relational DBMS. We call our method incremental query rewriting, because an original semantic query is rewritten into a (potentially infinite) series of SQL queries. In this chapter, we outline the main idea of our technique - using abstractions of databases and constrained clauses for deriving schematic answers, and provide completeness and soundness proofs to justify the applicability of this technique to the case of resolution for FOL without equality. The proposed method can be directly used with regular RDBs, including legacy databases. Moreover, we propose it as a potential basis for an efficient Web-scale semantic search technology.
Development of an Acoustic Signal Analysis Tool “Auto-F” Based on the Temperament Scale
NASA Astrophysics Data System (ADS)
Modegi, Toshio
The MIDI interface is originally designed for electronic musical instruments but we consider this music-note based coding concept can be extended for general acoustic signal description. We proposed applying the MIDI technology to coding of bio-medical auscultation sound signals such as heart sounds for retrieving medical records and performing telemedicine. Then we have tried to extend our encoding targets including vocal sounds, natural sounds and electronic bio-signals such as ECG, using Generalized Harmonic Analysis method. Currently, we are trying to separate vocal sounds included in popular songs and encode both vocal sounds and background instrumental sounds into separate MIDI channels. And also, we are trying to extract articulation parameters such as MIDI pitch-bend parameters in order to reproduce natural acoustic sounds using a GM-standard MIDI tone generator. In this paper, we present an overall algorithm of our developed acoustic signal analysis tool, based on those research works, which can analyze given time-based signals on the musical temperament scale. The prominent feature of this tool is producing high-precision MIDI codes, which reproduce the similar signals as the given source signal using a GM-standard MIDI tone generator, and also providing analyzed texts in the XML format.
Note on zero temperature holographic superfluids
NASA Astrophysics Data System (ADS)
Guo, Minyong; Lan, Shanquan; Niu, Chao; Tian, Yu; Zhang, Hongbao
2016-06-01
In this note, we have addressed various issues on zero temperature holographic superfluids. First, inspired by our numerical evidence for the equality between the superfluid density and particle density, we provide an elegant analytic proof for this equality by a boost trick. Second, using not only the frequency domain analysis but also the time domain analysis from numerical relativity, we identify the hydrodynamic normal modes and calculate out the sound speed, which is shown to increase with the chemical potential and saturate to the value predicted by the conformal field theory in the large chemical potential limit. Third, the generic non-thermalization is demonstrated by the fully nonlinear time evolution from a non-equilibrium state for our zero temperature holographic superfluid. Furthermore, a conserved Noether charge is proposed in support of this behavior.
Characteristic sounds facilitate visual search
Iordanescu, Lucica; Guzman-Martinez, Emmanuel; Grabowecky, Marcia; Suzuki, Satoru
2009-01-01
In a natural environment, objects that we look for often make characteristic sounds. A hiding cat may meow, or the keys in the cluttered drawer may jingle when moved. Using a visual search paradigm, we demonstrated that characteristic sounds facilitated visual localization of objects, even when the sounds carried no location information. For example, finding a cat was faster when participants heard a meow sound. In contrast, sounds had no effect when participants searched for names rather than pictures of objects. For example, hearing “meow” did not facilitate localization of the word cat. These results suggest that characteristic sounds cross-modally enhance visual (rather than conceptual) processing of the corresponding objects. Our behavioral demonstration of object-based cross-modal enhancement complements the extensive literature on space-based cross-modal interactions. When looking for your keys next time, you might want to play jingling sounds. PMID:18567253
Optical and Acoustic Sensor-Based 3D Ball Motion Estimation for Ball Sport Simulators †.
Seo, Sang-Woo; Kim, Myunggyu; Kim, Yejin
2018-04-25
Estimation of the motion of ball-shaped objects is essential for the operation of ball sport simulators. In this paper, we propose an estimation system for 3D ball motion, including speed and angle of projection, by using acoustic vector and infrared (IR) scanning sensors. Our system is comprised of three steps to estimate a ball motion: sound-based ball firing detection, sound source localization, and IR scanning for motion analysis. First, an impulsive sound classification based on the mel-frequency cepstrum and feed-forward neural network is introduced to detect the ball launch sound. An impulsive sound source localization using a 2D microelectromechanical system (MEMS) microphones and delay-and-sum beamforming is presented to estimate the firing position. The time and position of a ball in 3D space is determined from a high-speed infrared scanning method. Our experimental results demonstrate that the estimation of ball motion based on sound allows a wider activity area than similar camera-based methods. Thus, it can be practically applied to various simulations in sports such as soccer and baseball.
Relaxation phenomena in AOT-water-decane critical and dense microemulsions
NASA Astrophysics Data System (ADS)
Letamendia, L.; Pru-Lestret, E.; Panizza, P.; Rouch, J.; Sciortino, F.; Tartaglia, P.; Hashimoto, C.; Ushiki, H.; Risso, D.
2001-11-01
We report on extensive measurements of the low and high frequencies sound velocity and sound absorption in AOT-water-decane microemulsions deduced from ultrasonic and, for the first time as far as the absorption is concerned, from Brillouin scattering experiments. New experimental results on dielectric relaxation are also reported. Our results, which include data taken for critical as well as dense microemulsions, show new interesting relaxation phenomena. The relaxation frequencies deduced from very high frequency acoustical measurements are in good agreement with new high frequency dielectric relaxation measurements. We show that along the critical isochore, sound dispersion, relaxation frequency, and static dielectric permittivity can be accurately fitted to power laws. The absolute values of the new exponents we derived from experimental data are nearly equal, and they are very close to β=0.33 characterising the shape of the coexistence curve. The exponent characterising the infinite frequency permittivity is very close to 0.04 relevant to the diverging shear viscosity. For dense microemulsions, two well defined relaxation domains have been identified and the temperature variations of the sound absorption and the zero frequency dielectric permittivity bear striking similarities. We also show that the relaxation frequency of the slow relaxation process is almost independent of temperature and volume fraction and so cannot be attributed to percolation phenomena, whereas it can more likely be attributed to an intrinsic relaxation process probably connected to membrane fluctuations.
Augustine, Ann Mary; Chrysolyte, Shipra B; Thenmozhi, K; Rupa, V
2013-04-01
In order to assess psychosocial and auditory handicap in Indian patients with unilateral sensorineural hearing loss (USNHL), a prospective study was conducted on 50 adults with USNHL in the ENT Outpatient clinic of a tertiary care centre. The hearing handicap inventory for adults (HHIA) as well as speech in noise and sound localization tests were administered to patients with USNHL. An equal number of age-matched, normal controls also underwent the speech and sound localization tests. The results showed that HHIA scores ranged from 0 to 60 (mean 20.7). Most patients (84.8 %) had either mild to moderate or no handicap. Emotional subscale scores were higher than social subscale scores (p = 0.01). When the effect of sociodemographic factors on HHIA scores was analysed, educated individuals were found to have higher social subscale scores (p = 0.04). Age, sex, side and duration of hearing loss, occupation and income did not affect HHIA scores. Speech in noise and sound localization were significantly poorer in cases compared to controls (p < 0.001). About 75 % of patients refused a rehabilitative device. We conclude that USNHL in Indian adults does not usually produce severe handicap. When present, the handicap is more emotional than social. USNHL significantly affects sound localization and speech in noise. Yet, affected patients seldom seek a rehabilitative device.
Impact of NICU design on environmental noise.
Szymczak, Stacy E; Shellhaas, Renée A
2014-04-01
For neonates requiring intensive care, the optimal sound environment is uncertain. Minimal disruptions from medical staff create quieter environments for sleep, but limit language exposure necessary for proper language development. There are two models of neonatal intensive care units (NICUs): open-bay, in which 6-to-10 infants are cared for in a single large room; and single-room, in which neonates are housed in private, individual hospital rooms. We compared the acoustic environments in the two NICU models. We extracted the audio tracks from video-electroencephalography (EEG) monitoring studies from neonates in an open-bay NICU and compared the acoustic environment to that recorded from neonates in a new single-room NICU. From each NICU, 18 term infants were studied (total N=36; mean gestational age 39.3±1.9 weeks). Neither z-scores of the sound level variance (0.088±0.03 vs. 0.083±0.03, p=0.7), nor percent time with peak sound variance (above 2 standard deviations; 3.6% vs. 3.8%, p=0.6) were different. However, time below 0.05 standard deviations was higher in the single-room NICU (76% vs. 70%, p=0.02). We provide objective evidence that single-room NICUs have equal sound peaks and overall noise level variability compared with open-bay units, but the former may offer significantly more time at lower noise levels.
Atmospheric absorption of high frequency noise and application to fractional-octave bands
NASA Technical Reports Server (NTRS)
Shields, F. D.; Bass, H. E.
1977-01-01
Pure tone sound absorption coefficients were measured at 1/12 octave intervals from 4 to 100 KHz at 5.5K temperature intervals between 255.4 and 310.9 K and at 10 percent relative humidity increments between 0 percent and saturation in a large cylindrical tube (i.d., 25.4 cm; length, 4.8 m). Special solid-dielectric capacitance transducers, one to generate bursts of sound waves and one to terminate the sound path and detect the tone bursts, were constructed to fit inside the tube. The absorption was measured by varying the transmitter receiver separation from 1 to 4 m and observing the decay of multiple reflections or change in amplitude of the first received burst. The resulting absorption was compared with that from a proposed procedure for computing sound absorption in still air. Absorption of bands of noise was numerically computed by using the pure tone results. The results depended on spectrum shape, on filter type, and nonlinearly on propagation distance. For some of the cases considered, comparison with the extrapolation of ARP-866A showed a difference as large as a factor of 2. However, for many cases, the absorption for a finite band was nearly equal to the pure tone absorption at the center frequency of the band. A recommended prediction procedure is described for 1/3 octave band absorption coefficients.
Lee, Hyun-Ho; Lee, Sang-Kwon
2009-09-01
Booming sound is one of the important sounds in a passenger car. The aim of the paper is to develop the objective evaluation method of interior booming sound. The development method is based on the sound metrics and ANN (artificial neural network). The developed method is called the booming index. Previous work maintained that booming sound quality is related to loudness and sharpness--the sound metrics used in psychoacoustics--and that the booming index is developed by using the loudness and sharpness for a signal within whole frequency between 20 Hz and 20 kHz. In the present paper, the booming sound quality was found to be effectively related to the loudness at frequencies below 200 Hz; thus the booming index is updated by using the loudness of the signal filtered by the low pass filter at frequency under 200 Hz. The relationship between the booming index and sound metric is identified by an ANN. The updated booming index has been successfully applied to the objective evaluation of the booming sound quality of mass-produced passenger cars.
Semi-Supervised Active Learning for Sound Classification in Hybrid Learning Environments.
Han, Wenjing; Coutinho, Eduardo; Ruan, Huabin; Li, Haifeng; Schuller, Björn; Yu, Xiaojie; Zhu, Xuan
2016-01-01
Coping with scarcity of labeled data is a common problem in sound classification tasks. Approaches for classifying sounds are commonly based on supervised learning algorithms, which require labeled data which is often scarce and leads to models that do not generalize well. In this paper, we make an efficient combination of confidence-based Active Learning and Self-Training with the aim of minimizing the need for human annotation for sound classification model training. The proposed method pre-processes the instances that are ready for labeling by calculating their classifier confidence scores, and then delivers the candidates with lower scores to human annotators, and those with high scores are automatically labeled by the machine. We demonstrate the feasibility and efficacy of this method in two practical scenarios: pool-based and stream-based processing. Extensive experimental results indicate that our approach requires significantly less labeled instances to reach the same performance in both scenarios compared to Passive Learning, Active Learning and Self-Training. A reduction of 52.2% in human labeled instances is achieved in both of the pool-based and stream-based scenarios on a sound classification task considering 16,930 sound instances.
Semi-Supervised Active Learning for Sound Classification in Hybrid Learning Environments
Han, Wenjing; Coutinho, Eduardo; Li, Haifeng; Schuller, Björn; Yu, Xiaojie; Zhu, Xuan
2016-01-01
Coping with scarcity of labeled data is a common problem in sound classification tasks. Approaches for classifying sounds are commonly based on supervised learning algorithms, which require labeled data which is often scarce and leads to models that do not generalize well. In this paper, we make an efficient combination of confidence-based Active Learning and Self-Training with the aim of minimizing the need for human annotation for sound classification model training. The proposed method pre-processes the instances that are ready for labeling by calculating their classifier confidence scores, and then delivers the candidates with lower scores to human annotators, and those with high scores are automatically labeled by the machine. We demonstrate the feasibility and efficacy of this method in two practical scenarios: pool-based and stream-based processing. Extensive experimental results indicate that our approach requires significantly less labeled instances to reach the same performance in both scenarios compared to Passive Learning, Active Learning and Self-Training. A reduction of 52.2% in human labeled instances is achieved in both of the pool-based and stream-based scenarios on a sound classification task considering 16,930 sound instances. PMID:27627768
Mapping proteins in the presence of paralogs using units of coevolution
2013-01-01
Background We study the problem of mapping proteins between two protein families in the presence of paralogs. This problem occurs as a difficult subproblem in coevolution-based computational approaches for protein-protein interaction prediction. Results Similar to prior approaches, our method is based on the idea that coevolution implies equal rates of sequence evolution among the interacting proteins, and we provide a first attempt to quantify this notion in a formal statistical manner. We call the units that are central to this quantification scheme the units of coevolution. A unit consists of two mapped protein pairs and its score quantifies the coevolution of the pairs. This quantification allows us to provide a maximum likelihood formulation of the paralog mapping problem and to cast it into a binary quadratic programming formulation. Conclusion CUPID, our software tool based on a Lagrangian relaxation of this formulation, makes it, for the first time, possible to compute state-of-the-art quality pairings in a few minutes of runtime. In summary, we suggest a novel alternative to the earlier available approaches, which is statistically sound and computationally feasible. PMID:24564758
NASA Astrophysics Data System (ADS)
Diez, Matteo; Iemma, Umberto
2012-05-01
The article presents a novel approach to include community noise considerations based on sound quality in the Multidisciplinary Conceptual Design Optimization (MCDO) of civil transportation aircraft. The novelty stems from the use of an unconventional objective function, defined as a measure of the difference between the noise emission of the aircraft under analysis and a reference 'weakly annoying' noise, the target sound. The minimization of such a merit factor yields an aircraft concept with a noise signature as close as possible to the given target. The reference sound is one of the outcomes of the European Research Project SEFA (Sound Engineering For Aircraft, VI Framework Programme, 2004-2007), and used here as an external input. The aim of the present work is to address the definition and the inclusion of the sound-matching-based objective function in the MCDO of aircraft.
NASA Astrophysics Data System (ADS)
Nishiura, Takanobu; Nakamura, Satoshi
2003-10-01
Humans communicate with each other through speech by focusing on the target speech among environmental sounds in real acoustic environments. We can easily identify the target sound from other environmental sounds. For hands-free speech recognition, the identification of the target speech from environmental sounds is imperative. This mechanism may also be important for a self-moving robot to sense the acoustic environments and communicate with humans. Therefore, this paper first proposes hidden Markov model (HMM)-based environmental sound source identification. Environmental sounds are modeled by three states of HMMs and evaluated using 92 kinds of environmental sounds. The identification accuracy was 95.4%. This paper also proposes a new HMM composition method that composes speech HMMs and an HMM of categorized environmental sounds for robust environmental sound-added speech recognition. As a result of the evaluation experiments, we confirmed that the proposed HMM composition outperforms the conventional HMM composition with speech HMMs and a noise (environmental sound) HMM trained using noise periods prior to the target speech in a captured signal. [Work supported by Ministry of Public Management, Home Affairs, Posts and Telecommunications of Japan.
Transfer of knowledge from sound quality measurement to noise impact evaluation
NASA Astrophysics Data System (ADS)
Genuit, Klaus
2004-05-01
It is well known that the measurement and analysis of sound quality requires a complex procedure with consideration of the physical, psychoacoustical and psychological aspects of sound. Sound quality cannot be described only by a simple value based on A-weighted sound pressure level measurements. The A-weighted sound pressure level is sufficient to predict the probabilty that the human ear could be damaged by sound but the A-weighted level is not the correct descriptor for the annoyance of a complex sound situation given by several different sound events at different and especially moving positions (soundscape). On the one side, the consideration of the spectral distribution and the temporal pattern (psychoacoustics) is requested and, on the other side, the subjective attitude with respect to the sound situation, the expectation and experience of the people (psychology) have to be included in context with the complete noise impact evaluation. This paper describes applications of the newest methods of sound quality measurements-as it is well introduced at the car manufacturers-based on artifical head recordings and signal processing comparable to the human hearing used in noisy environments like community/traffic noise.
Environmental Sound Training in Cochlear Implant Users
Sheft, Stanley; Kuvadia, Sejal; Gygi, Brian
2015-01-01
Purpose The study investigated the effect of a short computer-based environmental sound training regimen on the perception of environmental sounds and speech in experienced cochlear implant (CI) patients. Method Fourteen CI patients with the average of 5 years of CI experience participated. The protocol consisted of 2 pretests, 1 week apart, followed by 4 environmental sound training sessions conducted on separate days in 1 week, and concluded with 2 posttest sessions, separated by another week without training. Each testing session included an environmental sound test, which consisted of 40 familiar everyday sounds, each represented by 4 different tokens, as well as the Consonant Nucleus Consonant (CNC) word test, and Revised Speech Perception in Noise (SPIN-R) sentence test. Results Environmental sounds scores were lower than for either of the speech tests. Following training, there was a significant average improvement of 15.8 points in environmental sound perception, which persisted 1 week later after training was discontinued. No significant improvements were observed for either speech test. Conclusions The findings demonstrate that environmental sound perception, which remains problematic even for experienced CI patients, can be improved with a home-based computer training regimen. Such computer-based training may thus provide an effective low-cost approach to rehabilitation for CI users, and potentially, other hearing impaired populations. PMID:25633579
Nothing Virtual About it: An Emerging Safe Haven for an Adaptive Enemy
2010-12-02
TERRORISTS May Allah bless you lions of the front, for by Allah, the fruits of your combined efforts-- sound, video, and text—are more severe for the...insurgencies, as demonstrated in the Soviet-Afghan War and many other conflicts exists and for many insurgencies perhaps is its operational center of...who have limited geographical safe havens, the Internet is becoming of equal or greater importance if not their center of gravity. The Birth of Al
2007-03-29
meaningful decisions about cleanup of contaminated sediments and hazardous waste disposal sites, accurately assess and implement control of...invertebrate species, and dietary benchmarks were set to the No Observed Adverse Effect Level (NOAEL) for the consumption of prey by black ducks , ospreys...prey equal to NOAEL assuming 100% of diet obtained from single prey. Footnoted references cited in Sample et al. 1996. 6 Black duck (DBDuck) NOAEL
The acoustic experimental investigation of counterrotating propeller configurations
NASA Technical Reports Server (NTRS)
Gazzaniga, John A.
1987-01-01
An experimental study of scale counterrotating propellers operating in an anechoic facility has been conducted. Various configurations of counterrotation for equal numbers of blades per disk have been tested along with single-rotation propellers, underscoring the fundamental acoustic differences between single and counterrotation propeller operation. In addition it is shown that, as the loading on the counterrotating system is increased, the overall sound-pressure level is also increased in both the disk plane and axial direction.
The physics of musical scales: Theory and experiment
NASA Astrophysics Data System (ADS)
Durfee, Dallin S.; Colton, John S.
2015-10-01
The theory of musical scales involves mathematical ratios, harmonic resonators, beats, and human perception and provides an interesting application of the physics of waves and sound. We first review the history and physics of musical scales, with an emphasis on four historically important scales: twelve-tone equal temperament, Pythagorean, quarter-comma meantone, and Ptolemaic just intonation. We then present an easy way for students and teachers to directly experience the qualities of different scales using MIDI synthesis.
Study of photonuclear muon interactions at Baksan underground scintillation telescope
NASA Technical Reports Server (NTRS)
Bakatanov, V. N.; Chudakov, A. E.; Dadykin, V. L.; Novoseltsev, Y. F.; Achkasov, V. M.; Semenov, A. M.; Stenkin, Y. V.
1985-01-01
The method of pion-muon-electron decays recording was used to distinguish between purely electron-photon and hadronic cascades, induced by high energy muons underground. At energy approx. 1 Tev a ratio of the number of hadronic to electromagnetic cascades was found equal 0.11 + or - .03 in agreement with expectation. But, at an energy approx. 4 Tev a sharp increase of this ratio was indicated though not statistically sound (0.52 + or - .13).
A Method for Optimizing Non-Axisymmetric Liners for Multimodal Sound Sources
NASA Technical Reports Server (NTRS)
Watson, W. R.; Jones, M. G.; Parrott, T. L.; Sobieski, J.
2002-01-01
Central processor unit times and memory requirements for a commonly used solver are compared to that of a state-of-the-art, parallel, sparse solver. The sparse solver is then used in conjunction with three constrained optimization methodologies to assess the relative merits of non-axisymmetric versus axisymmetric liner concepts for improving liner acoustic suppression. This assessment is performed with a multimodal noise source (with equal mode amplitudes and phases) in a finite-length rectangular duct without flow. The sparse solver is found to reduce memory requirements by a factor of five and central processing time by a factor of eleven when compared with the commonly used solver. Results show that the optimum impedance of the uniform liner is dominated by the least attenuated mode, whose attenuation is maximized by the Cremer optimum impedance. An optimized, four-segmented liner with impedance segments in a checkerboard arrangement is found to be inferior to an optimized spanwise segmented liner. This optimized spanwise segmented liner is shown to attenuate substantially more sound than the optimized uniform liner and tends to be more effective at the higher frequencies. The most important result of this study is the discovery that when optimized, a spanwise segmented liner with two segments gives attenuations equal to or substantially greater than an optimized axially segmented liner with the same number of segments.
Gustafson, Samantha; Pittman, Andrea; Fanning, Robert
2013-06-01
This tutorial demonstrates the effects of tubing length and coupling type (i.e., foam tip or personal earmold) on hearing threshold and real-ear-to-coupler difference (RECD) measures. Hearing thresholds from 0.25 kHz through 8 kHz are reported at various tubing lengths for 28 normal-hearing adults between the ages of 22 and 31 years. RECD values are reported for 14 of the adults. All measures were made with an insert earphone coupled to a standard foam tip and with an insert earphone coupled to each participant's personal earmold. Threshold and RECD measures obtained with a personal earmold were significantly different from those obtained with a foam tip on repeated measures analyses of variance. One-sample t tests showed these differences to vary systematically with increasing tubing length, with the largest average differences (7-8 dB) occurring at 4 kHz. This systematic examination demonstrates the equal and opposite effects of tubing length on threshold and acoustic measures. Specifically, as tubing length increased, sound pressure level in the ear canal decreased, affecting both hearing thresholds and the real-ear portion of the RECDs. This demonstration shows that when the same coupling method is used to obtain the hearing thresholds and RECD, equal and accurate estimates of real-ear sound pressure level are obtained.
Nair, Erika L; Sousa, Rhonda; Wannagot, Shannon
Guidelines established by the AAA currently recommend behavioral testing when fitting frequency modulated (FM) systems to individuals with cochlear implants (CIs). A protocol for completing electroacoustic measures has not yet been validated for personal FM systems or digital modulation (DM) systems coupled to CI sound processors. In response, some professionals have used or altered the AAA electroacoustic verification steps for fitting FM systems to hearing aids when fitting FM systems to CI sound processors. More recently steps were outlined in a proposed protocol. The purpose of this research is to review and compare the electroacoustic test measures outlined in a 2013 article by Schafer and colleagues in the Journal of the American Academy of Audiology titled "A Proposed Electroacoustic Test Protocol for Personal FM Receivers Coupled to Cochlear Implant Sound Processors" to the AAA electroacoustic verification steps for fitting FM systems to hearing aids when fitting DM systems to CI users. Electroacoustic measures were conducted on 71 CI sound processors and Phonak Roger DM systems using a proposed protocol and an adapted AAA protocol. Phonak's recommended default receiver gain setting was used for each CI sound processor manufacturer and adjusted if necessary to achieve transparency. Electroacoustic measures were conducted on Cochlear and Advanced Bionics (AB) sound processors. In this study, 28 Cochlear Nucleus 5/CP810 sound processors, 26 Cochlear Nucleus 6/CP910 sound processors, and 17 AB Naida CI Q70 sound processors were coupled in various combinations to Phonak Roger DM dedicated receivers (25 Phonak Roger 14 receivers-Cochlear dedicated receiver-and 9 Phonak Roger 17 receivers-AB dedicated receiver) and 20 Phonak Roger Inspiro transmitters. Employing both the AAA and the Schafer et al protocols, electroacoustic measurements were conducted with the Audioscan Verifit in a clinical setting on 71 CI sound processors and Phonak Roger DM systems to determine transparency and verify FM advantage, comparing speech inputs (65 dB SPL) in an effort to achieve equal outputs. If transparency was not achieved at Phonak's recommended default receiver gain, adjustments were made to the receiver gain. The integrity of the signal was monitored with the appropriate manufacturer's monitor earphones. Using the AAA hearing aid protocol, 50 of the 71 CI sound processors achieved transparency, and 59 of the 71 CI sound processors achieved transparency when using the proposed protocol at Phonak's recommended default receiver gain. After the receiver gain was adjusted, 3 of 21 CI sound processors still did not meet transparency using the AAA protocol, and 2 of 12 CI sound processors still did not meet transparency using the Schafer et al proposed protocol. Both protocols were shown to be effective in taking reliable electroacoustic measurements and demonstrate transparency. Both protocols are felt to be clinically feasible and to address the needs of populations that are unable to reliably report regarding the integrity of their personal DM systems. American Academy of Audiology
ERIC Educational Resources Information Center
Neeson, John F.; Austin, Stephen
1975-01-01
Describes a method for the measurement of the velocity of sound in various liquids based on the Raman-Nath theory of light-sound interaction. Utilizes an analog computer program to calculate the intensity of light scattered into various diffraction orders. (CP)
Decision making and preferences for acoustic signals in choice situations by female crickets.
Gabel, Eileen; Kuntze, Janine; Hennig, R Matthias
2015-08-01
Multiple attributes usually have to be assessed when choosing a mate. Efficient choice of the best mate is complicated if the available cues are not positively correlated, as is often the case during acoustic communication. Because of varying distances of signalers, a female may be confronted with signals of diverse quality at different intensities. Here, we examined how available cues are weighted for a decision by female crickets. Two songs with different temporal patterns and/or sound intensities were presented in a choice paradigm and compared with female responses from a no-choice test. When both patterns were presented at equal intensity, preference functions became wider in choice situations compared with a no-choice paradigm. When the stimuli in two-choice tests were presented at different intensities, this effect was counteracted as preference functions became narrower compared with choice tests using stimuli of equal intensity. The weighting of intensity differences depended on pattern quality and was therefore non-linear. A simple computational model based on pattern and intensity cues reliably predicted female decisions. A comparison of processing schemes suggested that the computations for pattern recognition and directionality are performed in a network with parallel topology. However, the computational flow of information corresponded to serial processing. © 2015. Published by The Company of Biologists Ltd.
Selective Listening Point Audio Based on Blind Signal Separation and Stereophonic Technology
NASA Astrophysics Data System (ADS)
Niwa, Kenta; Nishino, Takanori; Takeda, Kazuya
A sound field reproduction method is proposed that uses blind source separation and a head-related transfer function. In the proposed system, multichannel acoustic signals captured at distant microphones are decomposed to a set of location/signal pairs of virtual sound sources based on frequency-domain independent component analysis. After estimating the locations and the signals of the virtual sources by convolving the controlled acoustic transfer functions with each signal, the spatial sound is constructed at the selected point. In experiments, a sound field made by six sound sources is captured using 48 distant microphones and decomposed into sets of virtual sound sources. Since subjective evaluation shows no significant difference between natural and reconstructed sound when six virtual sources and are used, the effectiveness of the decomposing algorithm as well as the virtual source representation are confirmed.
De Angelis, Vittoria; De Martino, Federico; Moerel, Michelle; Santoro, Roberta; Hausfeld, Lars; Formisano, Elia
2017-11-13
Pitch is a perceptual attribute related to the fundamental frequency (or periodicity) of a sound. So far, the cortical processing of pitch has been investigated mostly using synthetic sounds. However, the complex harmonic structure of natural sounds may require different mechanisms for the extraction and analysis of pitch. This study investigated the neural representation of pitch in human auditory cortex using model-based encoding and decoding analyses of high field (7 T) functional magnetic resonance imaging (fMRI) data collected while participants listened to a wide range of real-life sounds. Specifically, we modeled the fMRI responses as a function of the sounds' perceived pitch height and salience (related to the fundamental frequency and the harmonic structure respectively), which we estimated with a computational algorithm of pitch extraction (de Cheveigné and Kawahara, 2002). First, using single-voxel fMRI encoding, we identified a pitch-coding region in the antero-lateral Heschl's gyrus (HG) and adjacent superior temporal gyrus (STG). In these regions, the pitch representation model combining height and salience predicted the fMRI responses comparatively better than other models of acoustic processing and, in the right hemisphere, better than pitch representations based on height/salience alone. Second, we assessed with model-based decoding that multi-voxel response patterns of the identified regions are more informative of perceived pitch than the remainder of the auditory cortex. Further multivariate analyses showed that complementing a multi-resolution spectro-temporal sound representation with pitch produces a small but significant improvement to the decoding of complex sounds from fMRI response patterns. In sum, this work extends model-based fMRI encoding and decoding methods - previously employed to examine the representation and processing of acoustic sound features in the human auditory system - to the representation and processing of a relevant perceptual attribute such as pitch. Taken together, the results of our model-based encoding and decoding analyses indicated that the pitch of complex real life sounds is extracted and processed in lateral HG/STG regions, at locations consistent with those indicated in several previous fMRI studies using synthetic sounds. Within these regions, pitch-related sound representations reflect the modulatory combination of height and the salience of the pitch percept. Copyright © 2017 Elsevier Inc. All rights reserved.
2007-03-29
Development of An Empirical Water Quality Model for Stormwater Based on Watershed Land Use in Puget Sound Valerie I. Cullinan, Christopher W. May...Systems Center, Bremerton, WA) Introduction The Sinclair and Dyes Inlet watershed is located on the west side of Puget Sound in Kitsap County...Washington, U.S.A. (Figure 1). The Puget Sound Naval Shipyard (PSNS), U.S Environmental Protection Agency (USEPA), the Washington State Department of
Adaptive Modeling of Details for Physically-Based Sound Synthesis and Propagation
2015-03-21
the interface that ensures the consistency and validity of the solution given by the two methods. Transfer functions are used to model two-way...release; distribution is unlimited. Adaptive modeling of details for physically-based sound synthesis and propagation The views, opinions and/or...Research Triangle Park, NC 27709-2211 Applied sciences, Adaptive modeling , Physcially-based, Sound synthesis, Propagation, Virtual world REPORT
Heart Sound Biometric System Based on Marginal Spectrum Analysis
Zhao, Zhidong; Shen, Qinqin; Ren, Fangqin
2013-01-01
This work presents a heart sound biometric system based on marginal spectrum analysis, which is a new feature extraction technique for identification purposes. This heart sound identification system is comprised of signal acquisition, pre-processing, feature extraction, training, and identification. Experiments on the selection of the optimal values for the system parameters are conducted. The results indicate that the new spectrum coefficients result in a significant increase in the recognition rate of 94.40% compared with that of the traditional Fourier spectrum (84.32%) based on a database of 280 heart sounds from 40 participants. PMID:23429515
NASA Astrophysics Data System (ADS)
Hemmatian, M.; Sedaghati, R.
2016-04-01
This study aims to investigate the effect of using magnetorheological elastomer (MRE)-based adaptive tuned vibration absorbers (ATVA) on the sound transmission in an elastic plate. Sound transmission loss (STL) of an elastic circular thin plate is analytically studied. The plate is excited by a plane acoustic wave as an incident sound and the displacement of the plate is calculated using corresponding mode shapes of the system for clamped boundary condition. Rayleigh integral approach is used to express the transmitted sound pressure in terms of the plate's displacement modal amplitude. In order to increase sound transmission loss of the plate, the MRE-based ATVA is considered. The basic idea is to be able to change the stiffness of the ATVA by varying magnetic field in order to reduce the transmitted acoustic energy of the host structure in a wide frequency range. Here, a MRE-based ATVA under the shear mode consisting of an oscillator mass, magnetic conductor, coils and MRE is investigated. In order to predict the viscoelastic characteristics of the field-dependent MRE based on the applied magnetic field, the double pole model is used. Finally, MRE-based ATVAs are integrated with the plate to absorb the plate energy with the aim of decreasing the transmitted sound power. Results show that plate with integrated MRE-based ATVAs suppresses the axisymmetric vibration of the plate and thus considerably improves the STL. Parametric studies on the influence of the position of MRE-based ATVAs and the effects of applied current on their performance are also presented.
Plastic modes of listening: affordance in constructed sound environments
NASA Astrophysics Data System (ADS)
Sjolin, Anders
This thesis is concerned with how the ecological approach to perception with the inclusion of listening modes, informs the creation of sound art installation, or more specifically as referred to in this thesis as constructed sound environments. The basis for the thesis has been a practiced based research where the aim and purpose of the written part of this PhD project has been to critically investigate the area of sound art, in order to map various approaches towards participating in and listening to a constructed sound environment. The main areas has been the notion of affordance as coined by James J. Gibson (1986), listening modes as coined by Pierre Schaeffer (1966) and further developed by Michel Chion (1994), aural architects as coined by Blesser and Salter (2007) and the holistic approach towards understanding sound art developed by Brandon LaBelle (2006). The findings within the written part of the thesis, based on a qualitative analysis, have informed the practice that has resulted in artefacts in the form of seven constructed sound environments that also functions as case studies for further analysis. The aim of the practice has been to exemplify the methodology, strategy and progress behind the organisation and construction of sound environments The research concerns point towards the acknowledgment of affordance as the crucial factor in understanding a constructed sound environment. The affordance approach govern the idea that perceiving a sound environment is a top-down process where the autonomic quality of a constructed sound environment is based upon the perception of structures of the sound material and its relationship with speaker placement and surrounding space. This enables a researcher to side step the conflicting poles of musical/abstract and non-musical/realistic classification of sound elements and regard these poles as included, not separated elements in the analysis of a constructed sound environment.
Using science soundly: The Yucca Mountain standard
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fri, R.W.
1995-09-01
Using sound science to shape government regulation is one of the most hotly argued topics in the ongoing debate about regulatory reform. Even though no one advaocates using unsound science, the belief that even the best science will sweep away regulatory controversy is equally foolish. As chair of a National Research Council (NRC) committee that studied the scientific basis for regulating high-level nuclear waste disposal, the author learned that science alone could resolve few of the key regulatory questions. Developing a standard that specifies a socially acceptable limit on the human health effects of nuclear waste releases involves many decisions.more » As the NRC committee learned in evaluating the scientific basis for the Yucca Mountain standard, a scientifically best decision rarely exists. More often, science can only offer a useful framework and starting point for policy debates. And sometimes, science`s most helpful contribution is to admit that it has nothing to say. The Yucca mountain study clearly illustrates that excessive faith in the power of science is more likely to produce messy frustration than crisp decisions. A better goal for regulatory reform is the sound use of science to clarify and contain the inevitable policy controversy.« less
Complementary satellite sound broadcasting systems: A NASA assessment for the Voice of America
NASA Technical Reports Server (NTRS)
Stevens, Grady H.; Spence, Rodney L.
1988-01-01
Satellite concepts are examined which offer potentially significant sound broadcast coverage of audio as a complement to VOA's existing and planned terrestrial sound broadcasting system. HF bands are emphasized but additional discussion is included for systems using higher frequencies. Low altitude satellites, shuttle altitude (275 km) and sun synchronous (about 1600 to 1800 km), would not be practical for international broadcasting since many satellites would be required for reliable and widespread coverage. Two concepts are discussed which would offer significant and practical broadcast coverage at HF. One, an 8-hr posigrade equatorial orbit, would offer about 1 hr of widespread, twice daily, coverage to three areas of the globe. The time of coverage is even greater when confined to densely populated areas only (2 to 3 hrs). Another orbit, the Apogee at Constant Time/Equatorial (ACE), provides the same coverage, but only once daily to each area. The latter orbit is highly elliptical, allowing insertion of a greater payload (more broadcast channels) with the existing launch capability. The ACE and 8-hr orbit concepts led to systems of about equal costs, with the ACE being slightly better.
Reverse amblyopia with atropine treatment.
Hainline, Bryan C; Sprunger, Derek C; Plager, David A; Neely, Daniel E; Guess, Matthew G
2009-01-01
Occlusion, pharmacologic pernalization and combined therapy have been documented in controlled studies to effectively treat amblyopia with few complications. However, there remain concerns about the effectiveness and complications when, as in this case, there are not standardized treatment protocols. A retrospective chart review of 133 consecutive patients in one community based ophthalmology practice treated for amblyopia was performed. Treatments evaluated were occlusion only, atropine penalization, and combination of occlusion and atropine. Reverse amblyopia was defined as having occured when the visual acuity of the sound eye was 3 LogMar units worse than visual acuity of the amblyopia eye after treatment. Improvement in vision after 6 months and 1 year of amblyopia therapy was similar among all three groups: 0.26 LogMar lines and 0.30 in the atropine group, 0.32 and 0.34 in the occlusion group, and 0.24 and 0.32 in the combined group. Eight (6%) patients demonstrated reverse amblyopia. The mean age of those who developed reverse amblyopia was 3.5 years, 1.5 years younger than the mean age of the study population, 7/8 had strabismic amblyopia, 6/8 were on daily atropine and had a mean refractive error of +4.77 diopters in the amblyopic eye and +5.06 diopters in the sound eye. Reverse amblyopia did not occur with occlusion only therapy. In this community based ophthalmology practice, atropine, patching, and combination therapy appear to be equally effective modalities to treat ambyopia. Highly hyperopic patients under 4 years of age with dense, strabismic amblyopia and on daily atropine appeared to be most at risk for development of reverse amblyopia.
Seafloor Geodesy using Wave Gliders to study Earthquake and Tsunami Hazards at Subduction Zones
NASA Astrophysics Data System (ADS)
Sathiakumar, S.; Barbot, S.; Hill, E.; Peng, D.; Zerucha, J.; Suhaimee, S.; Chia, G.; Salamena, G. G.; Syahailatua, A.
2016-12-01
Land-based GNSS networks are now in place to monitor most subduction zones of the world. These provide valuable information about the amount of geodetic strain accumulated in the region, which in turn gives insight into the seismic potential. However, it is usually impossible to resolve activity on the megathrust near the trench using land-based GNSS data alone, given typical signal-to-noise ratios. Ship-based seafloor geodesy is being used today to fill this observation gap. However, surveys using ships are very expensive, tedious and impractical due to the large areas to be covered. Instead of discrete missions using ships, continuous monitoring of the seafloor using autonomous marine robots would aid in understanding the tectonic setting of the seafloor better at a potentially lower cost, as well as help in designing better warning systems. Thus, we are developing seafloor geodesy capabilities using Wave Gliders, a new class of wave-propelled, persistent marine autonomous vehicle using a combination of acoustic and GNSS technologies. We use GNSS/INS to position the platform, and acoustic ranging to locate the seafloor. The GNSS/INS system to be integrated with the Wave Gliders has stringent requirements of low power, light weight, and high accuracy. All these factors are equally important due to limited power and space in the Wave Gliders and the need for highly accurate and precise measurements. With this hardware setup, a limiting factor is the accuracy of measurement of the sound velocity in the water column. We plan to obtain precise positioning of seafloor by exploring a measurement setup that minimizes uncertainties in sound velocity. This will be achieved by making fine-resolution measurements of the two-way travel time of the acoustic waves underwater using the Wave Gliders, and performing statistical signal processing on this data to obtain more reliable sound velocity measurement. This enhanced seafloor geodetic technique using Wave Gliders should enable us to measure deformation of the seafloor over annual and decadal time scales. Combination of seafloor data with land based GPS networks would allow us to study processes from the deep-ocean trench to the coast in a better way, and thereby give a more complete picture of the subduction process.
A laboratory study of the subjective response to helicopter blade-slap noise
NASA Technical Reports Server (NTRS)
Shepherd, K. P.
1978-01-01
The test stimuli recorded during a recent field study consisted of 16 sounds, each presented at 4 peak noise levels. Two helicopters and a fixed-wing aircraft were used. The impulsive characteristics of one helicopter were varied by operating at different rotor speeds, whereas the other helicopter, the noise of which was dominated by the tail rotor, displayed little variation in blade-slap noise. Thirty-two subjects made noisiness judgments on a continuous, 11 point, numerical scale. Preliminary results indicate that proposed impulsiveness corrections provide no significant improvement in the noisiness predictive ability of Effective Perceived Noise Levels (EPNL). For equal EPNL, the two categories of helicopter stimuli, one of which was far more impulsive than the other, showed no difference in judged noisiness. Examination of the physical characteristics of the sounds presented in the laboratory highlighted the difficulty of reproducing acoustical signals with high-crest factors.
Evaluation of the annoyance due to helicopter rotor noise
NASA Technical Reports Server (NTRS)
Sternfeld, H., Jr.; Doyle, L. B.
1978-01-01
A program was conducted in which 25 test subjects adjusted the levels of various helicopter rotor spectra until the combination of the harmonic noise and a broadband background noise was judged equally annoying as a higher level of the same broadband noise spectrum. The subjective measure of added harmonic noise was equated to the difference in the two levels of broadband noise. The test participants also made subjective evaluations of the rotor noise signatures which they created. The test stimuli consisted of three degrees of rotor impulsiveness, each presented at four blade passage rates. Each of these 12 harmonic sounds was combined with three broadband spectra and was adjusted to match the annoyance of three different sound pressure levels of broadband noise. Analysis of variance indicated that the important variables were level and impulsiveness. Regression analyses indicated that inclusion of crest factor improved correlation between the subjective measures and various objective or physical measures.
USAF Bioenvironmental Noise Data Handbook. Volume 165: MC-1 heater, duct type, portable
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-06-01
The MC-1 heater is a gasoline-motor driven, portable ground heater used primarily for cockpit and cabin temperature control. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at normal rated conditions. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
USAF bioenvironmental noise data handbook. Volume 158: F-106A aircraft, near and far-field noise
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The USAF F-106A is a single seat, all-weather fighter/interceptor aircraft powered by a J75-P-17 turbojet engine. This report provides measured and extrapolated data defining the bioacoustic environments produced by this aircraft operating on a concrete runup pad for five engine-power conditions. Near-field data are reported for five locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise levels, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 19 locations are normalized to standard meteorological conditions and extrapolated from 75 - 8000 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
USAF bioenvironmental noise data handbook. Volume 163: GPC-28 compressor
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The GPC-28 is a gasoline engine-driven compressor with a 120 volt 60 Hz generator used for general purpose maintenance. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at a normal rated condition. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 - 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
USAF bioenvironmental noise data handbook. Volume 161: A/M32A-86 generator set, diesel engine driven
NASA Astrophysics Data System (ADS)
Rau, T. H.
1982-05-01
The A/M32A-86 generator set is a diesel engine driven source of electrical power used for the starting of aircraft, and for ground maintenance. This report provides measured and extrapolated data defining the bioacoustic environments produced by this unit operating outdoors on a concrete apron at normal rated/loaded conditions. Near-field data are reported for 37 locations in a wide variety of physical and psychoacoustic measures: overall and band sound pressure levels, C-weighted and A-weighted sound levels, preferred speech interference level, perceived noise level, and limiting times for total daily exposure of personnel with and without standard Air Force ear protectors. Far-field data measured at 36 locations are normalized to standard meteorological conditions and extrapolated from 10 - 1600 meters to derive sets of equal-value contours for these same seven acoustic measures as functions of angle and distance from the source.
Active noise control: A tutorial for HVAC designers
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gelin, L.J.
1997-08-01
This article will identify the capabilities and limitations of ANC in its application to HVAC noise control. ANC can be used in ducted HVAC systems to cancel ductborne, low-frequency fan noise by injecting sound waves of equal amplitude and opposite phase into an air duct, as close as possible to the source of the unwanted noise. Destructive interference of the fan noise and injected noise results in sound cancellation. The noise problems that it solves are typically described as rumble, roar or throb, all of which are difficult to address using traditional noise control methods. This article will also contrastmore » the use of active against passive noise control techniques. The main differences between the two noise control measures are acoustic performance, energy consumption, and design flexibility. The article will first present the fundamentals and basic physics of ANC. The application to real HVAC systems will follow.« less
Steerable sound transport in a 3D acoustic network
NASA Astrophysics Data System (ADS)
Xia, Bai-Zhan; Jiao, Jun-Rui; Dai, Hong-Qing; Yin, Sheng-Wen; Zheng, Sheng-Jie; Liu, Ting-Ting; Chen, Ning; Yu, De-Jie
2017-10-01
Quasi-lossless and asymmetric sound transports, which are exceedingly desirable in various modern physical systems, are almost always based on nonlinear or angular momentum biasing effects with extremely high power levels and complex modulation schemes. A practical route for the steerable sound transport along any arbitrary acoustic pathway, especially in a three-dimensional (3D) acoustic network, can revolutionize the sound power propagation and the sound communication. Here, we design an acoustic device containing a regular-tetrahedral cavity with four cylindrical waveguides. A smaller regular-tetrahedral solid in this cavity is eccentrically emplaced to break spatial symmetry of the acoustic device. The numerical and experimental results show that the sound power flow can unimpededly transport between two waveguides away from the eccentric solid within a wide frequency range. Based on the quasi-lossless and asymmetric transport characteristic of the single acoustic device, we construct a 3D acoustic network, in which the sound power flow can flexibly propagate along arbitrary sound pathways defined by our acoustic devices with eccentrically emplaced regular-tetrahedral solids.
Artificial intelligence techniques used in respiratory sound analysis--a systematic review.
Palaniappan, Rajkumar; Sundaraj, Kenneth; Sundaraj, Sebastian
2014-02-01
Artificial intelligence (AI) has recently been established as an alternative method to many conventional methods. The implementation of AI techniques for respiratory sound analysis can assist medical professionals in the diagnosis of lung pathologies. This article highlights the importance of AI techniques in the implementation of computer-based respiratory sound analysis. Articles on computer-based respiratory sound analysis using AI techniques were identified by searches conducted on various electronic resources, such as the IEEE, Springer, Elsevier, PubMed, and ACM digital library databases. Brief descriptions of the types of respiratory sounds and their respective characteristics are provided. We then analyzed each of the previous studies to determine the specific respiratory sounds/pathology analyzed, the number of subjects, the signal processing method used, the AI techniques used, and the performance of the AI technique used in the analysis of respiratory sounds. A detailed description of each of these studies is provided. In conclusion, this article provides recommendations for further advancements in respiratory sound analysis.
NASA Astrophysics Data System (ADS)
Suzuki, Yôiti; Watanabe, Kanji; Iwaya, Yukio; Gyoba, Jiro; Takane, Shouichi
2005-04-01
Because the transfer functions governing subjective sound localization (HRTFs) show strong individuality, sound localization systems based on synthesis of HRTFs require suitable HRTFs for individual listeners. However, it is impractical to obtain HRTFs for all listeners based on measurements. Improving sound localization by adjusting non-individualized HRTFs to a specific listener based on that listener's anthropometry might be a practical method. This study first developed a new method to estimate interaural time differences (ITDs) using HRTFs. Then correlations between ITDs and anthropometric parameters were analyzed using the canonical correlation method. Results indicated that parameters relating to head size, and shoulder and ear positions are significant. Consequently, it was attempted to express ITDs based on listener's anthropometric data. In this process, the change of ITDs as a function of azimuth angle was parameterized as a sum of sine functions. Then the parameters were analyzed using multiple regression analysis, in which the anthropometric parameters were used as explanatory variables. The predicted or individualized ITDs were installed in the nonindividualized HRTFs to evaluate sound localization performance. Results showed that individualization of ITDs improved horizontal sound localization.
Opo lidar sounding of trace atmospheric gases in the 3 - 4 μm spectral range
NASA Astrophysics Data System (ADS)
Romanovskii, Oleg A.; Sadovnikov, Sergey A.; Kharchenko, Olga V.; Yakovlev, Semen V.
2018-04-01
The applicability of a KTA crystal-based laser system with optical parametric oscillators (OPO) generation to lidar sounding of the atmosphere in the spectral range 3-4 μm is studied in this work. A technique developed for lidar sounding of trace atmospheric gases (TAG) is based on differential absorption lidar (DIAL) method and differential optical absorption spectroscopy (DOAS). The DIAL-DOAS technique is tested to estimate its efficiency for lidar sounding of atmospheric trace gases. The numerical simulation performed shows that a KTA-based OPO laser is a promising source of radiation for remote DIAL-DOAS sounding of the TAGs under study along surface tropospheric paths. A possibility of using a PD38-03-PR photodiode for the DIAL gas analysis of the atmosphere is shown.
Nanocellulose based polymer composite for acoustical materials
NASA Astrophysics Data System (ADS)
Farid, Mohammad; Purniawan, Agung; Susanti, Diah; Priyono, Slamet; Ardhyananta, Hosta; Rahmasita, Mutia E.
2018-04-01
Natural fibers are biodegradable materials that are innovatively and widely used for composite reinforcement in automotive components. Nanocellulose derived from natural fibers oil palm empty bunches have properties that are remarkable for use as a composite reinforcement. However, there have not been many investigations related to the use of nanocellulose-based composites for wideband sound absorption materials. The specimens of nanocellulose-based polyester composite were prepared using a spray method. An impedance tube method was used to measure the sound absorption coefficient of this composite material. To reveal the characteristics of the nanocellulose-based polyester composite material, SEM (scanning electron microscope), TEM (Transmission Electron Microscope), FTIR (Fourier Transform Infra Red), TGA (Thermogravimetric Analysis), and density tests were performed. Sound absorption test results showed the average value of sound absorption coefficient of 0.36 to 0,46 for frequency between 500 and 4000 Hz indicating that this nanocellulose-based polyester composite materials had a tendency to wideband sound absorption materials and potentially used as automotive interior materials.
NASA Astrophysics Data System (ADS)
Gibson, J. Murray
2009-03-01
In looking at the commonalities between music and science, one sees that the musician's palette is based on the principles of physics. The pitch of a musical note is determined by the frequency of the sound wave. The scales that musicians use to create and play music can be viewed as a set of rules. What makes music interesting is how musicians develop those rules and create ambiguity with them. I will discuss the evolution of western musical scales in this context. As a particular example, ``Blue'' notes are very harmonic notes that are missing from the equal temperament scale. The techniques of piano blues and jazz represent the melding of African and Western music into something totally new and exciting. Live keyboard demonstrations will be used. Beyond any redeeming entertainment value the talk will emphasize the serious connections between science and art in music. Nevertheless tips will be accepted.
Li, Chenxi; Cazzolato, Ben; Zander, Anthony
2016-01-01
The classic analytical model for the sound absorption of micro perforated materials is well developed and is based on a boundary condition where the velocity of the material is assumed to be zero, which is accurate when the material vibration is negligible. This paper develops an analytical model for finite-sized circular micro perforated membranes (MPMs) by applying a boundary condition such that the velocity of air particles on the hole wall boundary is equal to the membrane vibration velocity (a zero-slip condition). The acoustic impedance of the perforation, which varies with its position, is investigated. A prediction method for the overall impedance of the holes and the combined impedance of the MPM is also provided. The experimental results for four different MPM configurations are used to validate the model and good agreement between the experimental and predicted results is achieved.
Algorithm of resonance orders for the objects
NASA Astrophysics Data System (ADS)
Zhang, YongGang; Zhang, JianXue
2018-03-01
In mechanical engineering, the object resonance phenomena often occur when the external incident wave frequency is close to object of the natural frequency. Object resonance phenomena get the maximum value when the external incident frequency is equal to object the natural frequency. Experiments found that resonance intension of the object is changed, different objects resonance phenomena present different characteristics of ladders. Based on object orders resonance characteristics, the calculation method of object orders resonance is put forward in the paper, and the application for the light and sound waves on the seven order resonance characteristics by people feel, the result error is less than 1%.Visible in this paper, the method has high accuracy and usability. The calculation method reveals that some object resonance occur present order characteristic only four types, namely the first-orders resonance characteristics, third-orders characteristics, five orders characteristic, and seven orders characteristic.
Modeling and Sound Insulation Performance Analysis of Two Honeycomb-hole Coatings
NASA Astrophysics Data System (ADS)
Ye, H. F.; Tao, M.; Zhang, W. Z.
2018-05-01
During the sound transmission loss test in the standing-wave tube, the unavoidable reflected wave from the termination of downstream tube would affect the precision measurement of sound transmission loss(TL). However, it can be solved by defining the non-reflected boundary conditions when modeling based on the finite element method. Then, the model has been validated by comparing with the analytical method. Based on the present model, the sound insulation performance of two types of honeycomb-hole coatings have been analyzed and discussed. Moreover, the changes of parameters play an important role on the sound insulation performance of honeycomb-hole coating and the negative Poisson’s ratio honeycomb-hole coating has better sound insulation performance at special frequencies. Finally, it is summarized that sound insulation performance is the result of various factors that include the impedance changes, the waveform transformation and so on.
Zhang, Lanyue; Ding, Dandan; Yang, Desen; Wang, Jia; Shi, Jie
2017-01-01
Spherical microphone arrays have been paid increasing attention for their ability to locate a sound source with arbitrary incident angle in three-dimensional space. Low-frequency sound sources are usually located by using spherical near-field acoustic holography. The reconstruction surface and holography surface are conformal surfaces in the conventional sound field transformation based on generalized Fourier transform. When the sound source is on the cylindrical surface, it is difficult to locate by using spherical surface conformal transform. The non-conformal sound field transformation by making a transfer matrix based on spherical harmonic wave decomposition is proposed in this paper, which can achieve the transformation of a spherical surface into a cylindrical surface by using spherical array data. The theoretical expressions of the proposed method are deduced, and the performance of the method is simulated. Moreover, the experiment of sound source localization by using a spherical array with randomly and uniformly distributed elements is carried out. Results show that the non-conformal surface sound field transformation from a spherical surface to a cylindrical surface is realized by using the proposed method. The localization deviation is around 0.01 m, and the resolution is around 0.3 m. The application of the spherical array is extended, and the localization ability of the spherical array is improved. PMID:28489065
Performance of an open-source heart sound segmentation algorithm on eight independent databases.
Liu, Chengyu; Springer, David; Clifford, Gari D
2017-08-01
Heart sound segmentation is a prerequisite step for the automatic analysis of heart sound signals, facilitating the subsequent identification and classification of pathological events. Recently, hidden Markov model-based algorithms have received increased interest due to their robustness in processing noisy recordings. In this study we aim to evaluate the performance of the recently published logistic regression based hidden semi-Markov model (HSMM) heart sound segmentation method, by using a wider variety of independently acquired data of varying quality. Firstly, we constructed a systematic evaluation scheme based on a new collection of heart sound databases, which we assembled for the PhysioNet/CinC Challenge 2016. This collection includes a total of more than 120 000 s of heart sounds recorded from 1297 subjects (including both healthy subjects and cardiovascular patients) and comprises eight independent heart sound databases sourced from multiple independent research groups around the world. Then, the HSMM-based segmentation method was evaluated using the assembled eight databases. The common evaluation metrics of sensitivity, specificity, accuracy, as well as the [Formula: see text] measure were used. In addition, the effect of varying the tolerance window for determining a correct segmentation was evaluated. The results confirm the high accuracy of the HSMM-based algorithm on a separate test dataset comprised of 102 306 heart sounds. An average [Formula: see text] score of 98.5% for segmenting S1 and systole intervals and 97.2% for segmenting S2 and diastole intervals were observed. The [Formula: see text] score was shown to increases with an increases in the tolerance window size, as expected. The high segmentation accuracy of the HSMM-based algorithm on a large database confirmed the algorithm's effectiveness. The described evaluation framework, combined with the largest collection of open access heart sound data, provides essential resources for evaluators who need to test their algorithms with realistic data and share reproducible results.
Using therapeutic sound with progressive audiologic tinnitus management.
Henry, James A; Zaugg, Tara L; Myers, Paula J; Schechter, Martin A
2008-09-01
Management of tinnitus generally involves educational counseling, stress reduction, and/or the use of therapeutic sound. This article focuses on therapeutic sound, which can involve three objectives: (a) producing a sense of relief from tinnitus-associated stress (using soothing sound); (b) passively diverting attention away from tinnitus by reducing contrast between tinnitus and the acoustic environment (using background sound); and (c) actively diverting attention away from tinnitus (using interesting sound). Each of these goals can be accomplished using three different types of sound-broadly categorized as environmental sound, music, and speech-resulting in nine combinations of uses of sound and types of sound to manage tinnitus. The authors explain the uses and types of sound, how they can be combined, and how the different combinations are used with Progressive Audiologic Tinnitus Management. They also describe how sound is used with other sound-based methods of tinnitus management (Tinnitus Masking, Tinnitus Retraining Therapy, and Neuromonics).
Efficient techniques for wave-based sound propagation in interactive applications
NASA Astrophysics Data System (ADS)
Mehra, Ravish
Sound propagation techniques model the effect of the environment on sound waves and predict their behavior from point of emission at the source to the final point of arrival at the listener. Sound is a pressure wave produced by mechanical vibration of a surface that propagates through a medium such as air or water, and the problem of sound propagation can be formulated mathematically as a second-order partial differential equation called the wave equation. Accurate techniques based on solving the wave equation, also called the wave-based techniques, are too expensive computationally and memory-wise. Therefore, these techniques face many challenges in terms of their applicability in interactive applications including sound propagation in large environments, time-varying source and listener directivity, and high simulation cost for mid-frequencies. In this dissertation, we propose a set of efficient wave-based sound propagation techniques that solve these three challenges and enable the use of wave-based sound propagation in interactive applications. Firstly, we propose a novel equivalent source technique for interactive wave-based sound propagation in large scenes spanning hundreds of meters. It is based on the equivalent source theory used for solving radiation and scattering problems in acoustics and electromagnetics. Instead of using a volumetric or surface-based approach, this technique takes an object-centric approach to sound propagation. The proposed equivalent source technique generates realistic acoustic effects and takes orders of magnitude less runtime memory compared to prior wave-based techniques. Secondly, we present an efficient framework for handling time-varying source and listener directivity for interactive wave-based sound propagation. The source directivity is represented as a linear combination of elementary spherical harmonic sources. This spherical harmonic-based representation of source directivity can support analytical, data-driven, rotating or time-varying directivity function at runtime. Unlike previous approaches, the listener directivity approach can be used to compute spatial audio (3D audio) for a moving, rotating listener at interactive rates. Lastly, we propose an efficient GPU-based time-domain solver for the wave equation that enables wave simulation up to the mid-frequency range in tens of minutes on a desktop computer. It is demonstrated that by carefully mapping all the components of the wave simulator to match the parallel processing capabilities of the graphics processors, significant improvement in performance can be achieved compared to the CPU-based simulators, while maintaining numerical accuracy. We validate these techniques with offline numerical simulations and measured data recorded in an outdoor scene. We present results of preliminary user evaluations conducted to study the impact of these techniques on user's immersion in virtual environment. We have integrated these techniques with the Half-Life 2 game engine, Oculus Rift head-mounted display, and Xbox game controller to enable users to experience high-quality acoustics effects and spatial audio in the virtual environment.
Advanced Energy Conversion Concept for Beamed-Energy Propulsion.
1987-08-21
pulsed simultaneously at 40 HL. the thruster will operate almost iMen&1 within the infra - sonIc regime. c) Asymmetric thrust considerations During low...LSD WAVES) - / TIME I-*- INFRA - SONIC I LEVITATIUN FREQUENCY Firm U-27 .?AmpUtude-laimpd macro-pulam .5 " I A. SIDE VIEW CUNICAL TIP ,?I , i...s equal local sta ic), and the sonic velocity. Since cool air can rush in only at the speed of sound, one would want to minimize the expanded plasma
Practical Aspects of Posttrauma Reconstruction With an Intramedullary Lengthening Nail
2017-01-01
Summary: Limb equalization using the Ilizarov method has evolved and adapted the use of internal lengthening devices. One of the newest devices, termed “PRECICE,” is a magnetically controlled telescoping nail. Complications such as pin site infection and skin irritation are eliminated. Despite trauma surgeons’ familiarity with intramedullary nailing, the Ilizarov method requires sound knowledge of deformity analysis and awareness of specific complications associated with distraction osteogenesis. This manuscript discusses some of the practical preoperative and intraoperative components of limb lengthening. PMID:28486286
2 × 2 Tables: a note on Campbell's recommendation.
Busing, F M T A; Weaver, B; Dubois, S
2016-04-15
For 2 × 2 tables, Egon Pearson's N - 1 chi-squared statistic is theoretically more sound than Karl Pearson's chi-squared statistic, and provides more accurate p values. Moreover, Egon Pearson's N - 1 chi-squared statistic is equal to the Mantel-Haenszel chi-squared statistic for a single 2 × 2 table, and as such, is often available in statistical software packages like SPSS, SAS, Stata, or R, which facilitates compliance with Ian Campbell's recommendations. Copyright © 2015 John Wiley & Sons, Ltd.
Experimental and Analytical Study of the Hydroacoustics of Propellers in Rigid Ducts
2006-07-01
5.3 at this loading condition indicate that the axial distribution of velocity is nearly uniform, and equal to the nominal value of 0.42 as set by the...moderate loading case produced the lowest level of radiated sound. The decrease in the mean axial velocity from the 0=0.42 condition corresponds to a...RtQ) X Axial coordinate p Fluid density W Azimuthal coordinate I\\ Acoustic wavelength c/f 50 Flow coefficient ( ui-p_ Blade loading coefficient ( \\ 1
Transition to motherhood and the self: measurement, stability, and change.
Ruble, D N; Brooks-Gunn, J; Fleming, A S; Fitzmaurice, G; Stangor, C; Deutsch, F
1990-03-01
Different ways of conceptualizing and measuring change in attitudes during transition to motherhood are examined. A series of analyses was performed on data from a cross-sectional sample (N = 667) and a smaller longitudinal sample (n = 48) to demonstrate sound psychometric properties for 2 new scales and to show construct comparability across different phases of childbearing. For Childbearing Attitudes Questionnaire, results demonstrated equality of covariance for 16 scales and comparability of structure and meaning of 4 higher order factors--identification with motherhood, social orientation, self-confidence, and negative aspects of giving birth. For Mothering Self-Definition Questionnaire, results demonstrated equality of covariance of 5 scales and comparability of structure and meaning of a single higher order factor, interpreted as reflecting positive feelings about one's mothering characteristics. Analyses of correlations and mean differences identified areas of change and stability.
The Noisiness of Low-Frequency One-Third Octave Bands of Noise. M.S. Thesis - Southampton Univ.
NASA Technical Reports Server (NTRS)
Lawton, B. W.
1975-01-01
This study examined the relative noisiness of low frequency one-third octave bands of noise bounded by the bands centered at 25 Hz and 200 Hz, with intensities ranging from 50 db sound pressure level (SPL) to 95 db SPL. The thirty-two subjects used a method-of-adjustment technique, producing comparison-band intensities as noisy as standard bands centered at 100 Hz and 200 Hz with intensities of 60 db SPL and 72 db SPL. Four contours of equal noisiness were developed for one-third octave bands, extending down to 25 Hz and ranging in intensity from approximately 58 db SPL to 86 db SPL. These curves were compared with the contours of equal noisiness of Kryter and Pearsons. In the region of overlap (between 50 Hz and 200 Hz) the agreement was good.
24 CFR 51.106 - Implementation.
Code of Federal Regulations, 2010 CFR
2010-04-01
... day-night average sound level data are not available may be evaluated from NEF or CNEL analyses using.... The day-night average sound level may be estimated from the design hour L10 or Leq values by the.... The Department of Defense uses day-night average sound level based on C-weighted sound level...
Farid-ul-Hasnain, Syed; Johansson, Eva; Mogren, Ingrid; Krantz, Gunilla
2012-01-01
The purpose of this qualitative study was to explore perceptions and expectations of young males and females, in Karachi, Pakistan, regarding their life prospects and gender roles, with resulting implications for health behaviour. The main theme emerging was “Young adults’ prospects in life are hampered by psychosocial and gender equality constraints”. Gender inequality and the low status of women in society were described as major obstacles to the prosperity and development. Persistent withholding of information to the younger generation on sexual and reproductive health issues was perceived to increase exposure to health risks, particularly sexually transmitted infections (STIs). The present study reveals new discourses on equality among young adults, pointing towards an increasing, sound interaction between the sexes and aspirations for more gender equal relationships. The study further reveals serious misconceptions about HIV/AIDS. Such views and awareness among the younger generation constitutes a strong force towards change of traditional norms, including reproductive health behaviour, and calls for policy change. PMID:22980235
Zoos in the twenty-first century: Can't we find a better way to love nature?
NASA Astrophysics Data System (ADS)
Dewey-Platt, Lauren Kay
As a new millennium approaches, many forms of life on the planet and the environments in which they have evolved are increasingly threatened by human activities Wildlife is being marginalized, and native habitats are disappearing at an unprecedented rate. Equally disturbing is the impending demise of traditional human societies---peoples who have evolved outside of the confines and conditions of modern Western influence. The loss of these human and animal societies has occurred so rapidly that implications are largely unknown. Research on how modern Americans relate to animals, particularly wildlife, revealed a clear and disturbing incongruity best exemplified in the current paradigm of zoo exhibition and education. Although zoos purport to educate visitors about the ecology of natural environments and the universal plight of wildlife, research shows that people, particularly children, learn less about ecological principles in zoos with live animals than they do in non-living natural history exhibits. While designers employ a variety of visual techniques in natural history exhibition, environmental sound as an educational exhibit component is largely nonexistent. Many animal species communicate through sound, especially species in underwater environments. As the audio equivalent of a landscape, the soundscape is as important as any other habitat feature to the well-being of wildlife populations. Using recorded sounds of natural environments, an exhibition soundscape was designed and produced for Oceanario de Lisboa in Lisbon, Portugal---the centerpiece of the 1998 World Exposition in Lisbon, Portugal. With programmed sound serving as a major component of natural history exhibition, a conceptual design of a novel zoo for the twenty-first century was described. The "NewZew" concept is based on a growing awareness that the best way to save species is to salvage, preserve, and restore their natural habitats---activities that are largely antithetical to current zoo practice. In introducing visitors to the lifeways of traditional societies and employing recent and emerging technologies, the NewZew educates its visitors about the wonders of life on earth without displaying live animals.
Allen, Paul D.; Ison, James R.
2010-01-01
Auditory spatial acuity was measured in mice using prepulse inhibition (PPI) of the acoustic startle reflex (ASR) as the indicator response for stimulus detection. The prepulse was a “speaker swap” (SSwap), shifting a noise between two speakers located along the azimuth. Their angular separation, and the spectral composition and sound level of the noise were varied, as was the interstimulus interval (ISI) between SSwap and ASR elicitation. In Experiment 1 a 180° SSwap of wide band noise (WBN) was compared with WBN Onset and Offset. SSwap and WBN Onset had near equal effects, but less than Offset. In Experiment 2 WBN SSwap was measured with speaker separations of 15°, 22.5°, 45°, and 90°. Asymptotic level and the growth rate of PPI increased with increased separation from 15° to 90°, but even the 15° SSwap provided significant PPI for the mean performance of the group. SSwap in Experiment 3 used octave band noise (2–4, 4–8, 8–16, or 16–32 kHz) and separations of 7.5° to 180°. SSwap was most effective for the highest frequencies, with no significant PPI for SSwap below 8–16 kHz, or for separations of 7.5°. In Experiment 4 SSwap had WBN sound levels from 40 to 78 dB SPL, and separations of 22.5°, 45°, 90° and 180°: PPI increased with level, this effect varying with ISI and angular separation. These experiments extend the prior findings on sound localization in mice, and the dependence of PPI on ISI adds a reaction-time-like dimension to this behavioral analysis. PMID:20364886
DeGiuli, Eric; Laversanne-Finot, Adrien; Düring, Gustavo; Lerner, Edan; Wyart, Matthieu
2014-08-14
Connectedness and applied stress strongly affect elasticity in solids. In various amorphous materials, mechanical stability can be lost either by reducing connectedness or by increasing pressure. We present an effective medium theory of elasticity that extends previous approaches by incorporating the effect of compression, of amplitude e, allowing one to describe quantitative features of sound propagation, transport, the boson peak, and elastic moduli near the elastic instability occurring at a compression ec. The theory disentangles several frequencies characterizing the vibrational spectrum: the onset frequency where strongly-scattered modes appear in the vibrational spectrum, the pressure-independent frequency ω* where the density of states displays a plateau, the boson peak frequency ωBP found to scale as , and the Ioffe-Regel frequency ωIR where scattering length and wavelength become equal. We predict that sound attenuation crosses over from ω(4) to ω(2) behaviour at ω0, consistent with observations in glasses. We predict that a frequency-dependent length scale ls(ω) and speed of sound ν(ω) characterize vibrational modes, and could be extracted from scattering data. One key result is the prediction of a flat diffusivity above ω0, in agreement with previously unexplained observations. We find that the shear modulus does not vanish at the elastic instability, but drops by a factor of 2. We check our predictions in packings of soft particles and study the case of covalent networks and silica, for which we predict ωIR ≈ ωBP. Overall, our approach unifies sound attenuation, transport and length scales entering elasticity in a single framework where disorder is not the main parameter controlling the boson peak, in agreement with observations. This framework leads to a phase diagram where various glasses can be placed, connecting microscopic structure to vibrational properties.
Lucke, Roy E; Raub, Richard A; Thunder, Thomas E
2004-01-01
The automated wayside horn system is designed to replace the train horn as a means of alerting motorists to danger and thus enhancing safety at highway-rail grade crossings. Furthermore, the wayside horn directionality is such that the warning sound is broadcast over a smaller sector than the train horn, thereby reducing residential noise. This article examines the results of an evaluation comparing train horns with wayside horns in the village of Mundelein, Illinois, USA. The study derived from previous work in Gering, Nebraska, and Ames, Iowa.During the 3 months covering the 'before' (train horn) period and through to 'after' (wayside horn), more than 19 500 crossing gate closures were recorded on videotape at three crossings. Analysis showed motorist violation of level-crossing laws decreased 68%, from an average rate of 3.53 per 100 gate closings when train horns were in use to 1.12 per 100 with the wayside horn. The decrease was statistically significant. Of equal importance was the decrease in residential noise. Sound measurements taken in a sample of residential yards showed a decrease in sound levels by more than 10 decibels (dB) at most locations. When plotted as sound contours, decreases in the area of coverage ranged from 85% at the 90dB level to 65% at the 70dB level.However, there are two issues with the use of wayside horns that need to be resolved. First, and most important, is that the wayside horn starts sounding when the warning lights begin to flash. This startles motorists, and some stop on the rail tracks. A second issue is the frequent unwarranted activation of the system, which encourages people to ignore the gate.
Neural correlates of audiotactile phonetic processing in early-blind readers: an fMRI study.
Pishnamazi, Morteza; Nojaba, Yasaman; Ganjgahi, Habib; Amousoltani, Asie; Oghabian, Mohammad Ali
2016-05-01
Reading is a multisensory function that relies on arbitrary associations between auditory speech sounds and symbols from a second modality. Studies of bimodal phonetic perception have mostly investigated the integration of visual letters and speech sounds. Blind readers perform an analogous task by using tactile Braille letters instead of visual letters. The neural underpinnings of audiotactile phonetic processing have not been studied before. We used functional magnetic resonance imaging to reveal the neural correlates of audiotactile phonetic processing in 16 early-blind Braille readers. Braille letters and corresponding speech sounds were presented in unimodal, and congruent/incongruent bimodal configurations. We also used a behavioral task to measure the speed of blind readers in identifying letters presented via tactile and/or auditory modalities. Reaction times for tactile stimuli were faster. The reaction times for bimodal stimuli were equal to those for the slower auditory-only stimuli. fMRI analyses revealed the convergence of unimodal auditory and unimodal tactile responses in areas of the right precentral gyrus and bilateral crus I of the cerebellum. The left and right planum temporale fulfilled the 'max criterion' for bimodal integration, but activities of these areas were not sensitive to the phonetical congruency between sounds and Braille letters. Nevertheless, congruency effects were found in regions of frontal lobe and cerebellum. Our findings suggest that, unlike sighted readers who are assumed to have amodal phonetic representations, blind readers probably process letters and sounds separately. We discuss that this distinction might be due to mal-development of multisensory neural circuits in early blinds or it might be due to inherent differences between Braille and print reading mechanisms.
WODA Technical Guidance on Underwater Sound from Dredging.
Thomsen, Frank; Borsani, Fabrizio; Clarke, Douglas; de Jong, Christ; de Wit, Pim; Goethals, Fredrik; Holtkamp, Martine; Martin, Elena San; Spadaro, Philip; van Raalte, Gerard; Victor, George Yesu Vedha; Jensen, Anders
2016-01-01
The World Organization of Dredging Associations (WODA) has identified underwater sound as an environmental issue that needs further consideration. A WODA Expert Group on Underwater Sound (WEGUS) prepared a guidance paper in 2013 on dredging sound, including a summary of potential impacts on aquatic biota and advice on underwater sound monitoring procedures. The paper follows a risk-based approach and provides guidance for standardization of acoustic terminology and methods for data collection and analysis. Furthermore, the literature on dredging-related sounds and the effects of dredging sounds on marine life is surveyed and guidance on the management of dredging-related sound risks is provided.
Quality and loudness judgments for music subjected to compression limiting.
Croghan, Naomi B H; Arehart, Kathryn H; Kates, James M
2012-08-01
Dynamic-range compression (DRC) is used in the music industry to maximize loudness. The amount of compression applied to commercial recordings has increased over time due to a motivating perspective that louder music is always preferred. In contrast to this viewpoint, artists and consumers have argued that using large amounts of DRC negatively affects the quality of music. However, little research evidence has supported the claims of either position. The present study investigated how DRC affects the perceived loudness and sound quality of recorded music. Rock and classical music samples were peak-normalized and then processed using different amounts of DRC. Normal-hearing listeners rated the processed and unprocessed samples on overall loudness, dynamic range, pleasantness, and preference, using a scaled paired-comparison procedure in two conditions: un-equalized, in which the loudness of the music samples varied, and loudness-equalized, in which loudness differences were minimized. Results indicated that a small amount of compression was preferred in the un-equalized condition, but the highest levels of compression were generally detrimental to quality, whether loudness was equalized or varied. These findings are contrary to the "louder is better" mentality in the music industry and suggest that more conservative use of DRC may be preferred for commercial music.
Human Response to Simulated Low-Intensity Sonic Booms
NASA Technical Reports Server (NTRS)
Sullivan, Brenda M.
2004-01-01
NASA's High Speed Research (HSR ) program in the 1990s was intended to develop a technology base for a future High-Speed Civil Transport (HSCT). As part of this program, the NASA Langley Research Center sonic boom simulator (SBS) was built and used for a series of tests on subjective response to sonic booms. At the end of the HSR program, an HSCT was deemed impractical, but since then interest in supersonic flight has reawakened, this time focusing on a smaller aircraft suitable for a business jet. To respond to this interest, the Langley sonic boom simulator has been refurbished. The upgraded computer-controlled playback system is based on an SGI O2 computer, in place of the previous DEC MicroVAX. As the frequency response of the booth is not flat, an equalization filter is required. Because of the changes made during the renovation (new loudspeakers), the previous equalization filter no longer performed as well as before, so a new equalization filter has been designed. Booms to be presented in the booth are preprocessed using the filter. When the preprocessed signals are presented into the booth and measured with a microphone, the results are very similar to the intended shapes. Signals with short rise times and sharp "corners" are observed to have a small amount of "ringing" in the response. During the HSR program a considerable number of subjective tests were completed in the SBS. A summary of that research is given in Leatherwood et al. (Individual reports are available at http://techreports.larc.nasa.gov/ltrs/ltrs.html.) Topics of study included shaped sonic booms, asymmetrical booms, realistic (recorded) boom waveforms, indoor and outdoor booms shapes, among other factors. One conclusion of that research was that a loudness metric, like the Stevens Perceived Level (PL), predicted human reaction much more accurately than overpressure or unweighted sound pressure level. Structural vibration and rattle were not included in these studies.
Absolute auditory threshold: testing the absolute.
Heil, Peter; Matysiak, Artur
2017-11-02
The mechanisms underlying the detection of sounds in quiet, one of the simplest tasks for auditory systems, are debated. Several models proposed to explain the threshold for sounds in quiet and its dependence on sound parameters include a minimum sound intensity ('hard threshold'), below which sound has no effect on the ear. Also, many models are based on the assumption that threshold is mediated by integration of a neural response proportional to sound intensity. Here, we test these ideas. Using an adaptive forced choice procedure, we obtained thresholds of 95 normal-hearing human ears for 18 tones (3.125 kHz carrier) in quiet, each with a different temporal amplitude envelope. Grand-mean thresholds and standard deviations were well described by a probabilistic model according to which sensory events are generated by a Poisson point process with a low rate in the absence, and higher, time-varying rates in the presence, of stimulation. The subject actively evaluates the process and bases the decision on the number of events observed. The sound-driven rate of events is proportional to the temporal amplitude envelope of the bandpass-filtered sound raised to an exponent. We find no evidence for a hard threshold: When the model is extended to include such a threshold, the fit does not improve. Furthermore, we find an exponent of 3, consistent with our previous studies and further challenging models that are based on the assumption of the integration of a neural response that, at threshold sound levels, is directly proportional to sound amplitude or intensity. © 2017 Federation of European Neuroscience Societies and John Wiley & Sons Ltd.
Olsen, Kirk N; Dean, Roger T; Leung, Yvonne
2016-01-01
Phrasing facilitates the organization of auditory information and is central to speech and music. Not surprisingly, aspects of changing intensity, rhythm, and pitch are key determinants of musical phrases and their boundaries in instrumental note-based music. Different kinds of speech (such as tone- vs. stress-languages) share these features in different proportions and form an instructive comparison. However, little is known about whether or how musical phrasing is perceived in sound-based music, where the basic musical unit from which a piece is created is commonly non-instrumental continuous sounds, rather than instrumental discontinuous notes. This issue forms the target of the present paper. Twenty participants (17 untrained in music) were presented with six stimuli derived from sound-based music, note-based music, and environmental sound. Their task was to indicate each occurrence of a perceived phrase and qualitatively describe key characteristics of the stimulus associated with each phrase response. It was hypothesized that sound-based music does elicit phrase perception, and that this is primarily associated with temporal changes in intensity and timbre, rather than rhythm and pitch. Results supported this hypothesis. Qualitative analysis of participant descriptions showed that for sound-based music, the majority of perceived phrases were associated with intensity or timbral change. For the note-based piano piece, rhythm was the main theme associated with perceived musical phrasing. We modeled the occurrence in time of perceived musical phrases with recurrent event 'hazard' analyses using time-series data representing acoustic predictors associated with intensity, spectral flatness, and rhythmic density. Acoustic intensity and timbre (represented here by spectral flatness) were strong predictors of perceived musical phrasing in sound-based music, and rhythm was only predictive for the piano piece. A further analysis including five additional spectral measures linked to timbre strengthened the models. Overall, results show that even when little of the pitch and rhythm information important for phrasing in note-based music is available, phrasing is still perceived, primarily in response to changes of intensity and timbre. Implications for electroacoustic music composition and music recommender systems are discussed.
Olsen, Kirk N.; Dean, Roger T.; Leung, Yvonne
2016-01-01
Phrasing facilitates the organization of auditory information and is central to speech and music. Not surprisingly, aspects of changing intensity, rhythm, and pitch are key determinants of musical phrases and their boundaries in instrumental note-based music. Different kinds of speech (such as tone- vs. stress-languages) share these features in different proportions and form an instructive comparison. However, little is known about whether or how musical phrasing is perceived in sound-based music, where the basic musical unit from which a piece is created is commonly non-instrumental continuous sounds, rather than instrumental discontinuous notes. This issue forms the target of the present paper. Twenty participants (17 untrained in music) were presented with six stimuli derived from sound-based music, note-based music, and environmental sound. Their task was to indicate each occurrence of a perceived phrase and qualitatively describe key characteristics of the stimulus associated with each phrase response. It was hypothesized that sound-based music does elicit phrase perception, and that this is primarily associated with temporal changes in intensity and timbre, rather than rhythm and pitch. Results supported this hypothesis. Qualitative analysis of participant descriptions showed that for sound-based music, the majority of perceived phrases were associated with intensity or timbral change. For the note-based piano piece, rhythm was the main theme associated with perceived musical phrasing. We modeled the occurrence in time of perceived musical phrases with recurrent event ‘hazard’ analyses using time-series data representing acoustic predictors associated with intensity, spectral flatness, and rhythmic density. Acoustic intensity and timbre (represented here by spectral flatness) were strong predictors of perceived musical phrasing in sound-based music, and rhythm was only predictive for the piano piece. A further analysis including five additional spectral measures linked to timbre strengthened the models. Overall, results show that even when little of the pitch and rhythm information important for phrasing in note-based music is available, phrasing is still perceived, primarily in response to changes of intensity and timbre. Implications for electroacoustic music composition and music recommender systems are discussed. PMID:27997625
Kuriki, Shinya; Kobayashi, Yusuke; Kobayashi, Takanari; Tanaka, Keita; Uchikawa, Yoshinori
2013-02-01
The auditory steady-state response (ASSR) is a weak potential or magnetic response elicited by periodic acoustic stimuli with a maximum response at about a 40-Hz periodicity. In most previous studies using amplitude-modulated (AM) tones of stimulus sound, long lasting tones of more than 10 s in length were used. However, characteristics of the ASSR elicited by short AM tones have remained unclear. In this study, we examined magnetoencephalographic (MEG) ASSR using a sequence of sinusoidal AM tones of 0.78 s in length with various tone frequencies of 440-990 Hz in about one octave variation. It was found that the amplitude of the ASSR was invariant with tone frequencies when the level of sound pressure was adjusted along an equal-loudness curve. The amplitude also did not depend on the existence of preceding tone or difference in frequency of the preceding tone. When the sound level of AM tones was changed with tone frequencies in the same range of 440-990 Hz, the amplitude of ASSR varied in a proportional manner to the sound level. These characteristics are favorable for the use of ASSR in studying temporal processing of auditory information in the auditory cortex. The lack of adaptation in the ASSR elicited by a sequence of short tones may be ascribed to the neural activity of widely accepted generator of magnetic ASSR in the primary auditory cortex. Copyright © 2012 Elsevier B.V. All rights reserved.
Brunner, Monika; Stellzig-Eisenhauer, Angelika; Pröschel, Ute; Verres, Rolf; Komposch, Gerda
2005-11-01
To evaluate the immediate, long-term, and carry-over effects of nasopharyngoscopic biofeedback therapy in patients with cleft palate who exhibit velopharyngeal dysfunction (VPD). Pre- versus posttreatment and follow-up comparisons. Cleft palate center of the Heidelberg University Hospital, Heidelberg, Germany. Eleven patients with VPD who had received conventional speech therapy without showing significant improvement. A four-stage feedback procedure. The patients watched and evaluated their velopharyngeal (VP) valving during speech by an endoscopic image displayed on a video monitor. Two feedback sessions took place for every target sound. Mean occurrence of VP closure during speech sound production on different linguistic levels. Patients' self-perception was assessed by a questionnaire and speech diary. Significant improvement and stability of VP closure was noted. Mean occurrence of VP closure was 5% before therapy, 91% after two biofeedback sessions, and 86% in the follow-up after 6 months. Velopharyngeal dysfunction associated with compensatory articulation proved to be equally well trained as VPD on sounds with good articulatory placement. No significant difference was observed in the degree of improvement between phoneme-specific VPD and generalized VPD. The transfer to the level of words and sentences was successful and showed significant stability. The stability of VP closure for vowels was less than the stability for fricatives and stop sounds. Patients gained improved auditory and kinesthetic self-perception of their articulation. Nasopharyngoscopic biofeedback therapy proves to be a quick and effective method to change VPD. It shows stable results and carry-over effects.
Analysis of swallowing sounds using hidden Markov models.
Aboofazeli, Mohammad; Moussavi, Zahra
2008-04-01
In recent years, acoustical analysis of the swallowing mechanism has received considerable attention due to its diagnostic potentials. This paper presents a hidden Markov model (HMM) based method for the swallowing sound segmentation and classification. Swallowing sound signals of 15 healthy and 11 dysphagic subjects were studied. The signals were divided into sequences of 25 ms segments each of which were represented by seven features. The sequences of features were modeled by HMMs. Trained HMMs were used for segmentation of the swallowing sounds into three distinct phases, i.e., initial quiet period, initial discrete sounds (IDS) and bolus transit sounds (BTS). Among the seven features, accuracy of segmentation by the HMM based on multi-scale product of wavelet coefficients was higher than that of the other HMMs and the linear prediction coefficient (LPC)-based HMM showed the weakest performance. In addition, HMMs were used for classification of the swallowing sounds of healthy subjects and dysphagic patients. Classification accuracy of different HMM configurations was investigated. When we increased the number of states of the HMMs from 4 to 8, the classification error gradually decreased. In most cases, classification error for N=9 was higher than that of N=8. Among the seven features used, root mean square (RMS) and waveform fractal dimension (WFD) showed the best performance in the HMM-based classification of swallowing sounds. When the sequences of the features of IDS segment were modeled separately, the accuracy reached up to 85.5%. As a second stage classification, a screening algorithm was used which correctly classified all the subjects but one healthy subject when RMS was used as characteristic feature of the swallowing sounds and the number of states was set to N=8.
[Synchronous playing and acquiring of heart sounds and electrocardiogram based on labVIEW].
Dan, Chunmei; He, Wei; Zhou, Jing; Que, Xiaosheng
2008-12-01
In this paper is described a comprehensive system, which can acquire heart sounds and electrocardiogram (ECG) in parallel, synchronize the display; and play of heart sound and make auscultation and check phonocardiogram to tie in. The hardware system with C8051F340 as the core acquires the heart sound and ECG synchronously, and then sends them to indicators, respectively. Heart sounds are displayed and played simultaneously by controlling the moment of writing to indicator and sound output device. In clinical testing, heart sounds can be successfully located with ECG and real-time played.
ERIC Educational Resources Information Center
Amrani, D.
2013-01-01
This paper deals with the comparison of sound speed measurements in air using two types of sensor that are widely employed in physics and engineering education, namely a pressure sensor and a sound sensor. A computer-based laboratory with pressure and sound sensors was used to carry out measurements of air through a 60 ml syringe. The fast Fourier…
Series expansions of rotating two and three dimensional sound fields.
Poletti, M A
2010-12-01
The cylindrical and spherical harmonic expansions of oscillating sound fields rotating at a constant rate are derived. These expansions are a generalized form of the stationary sound field expansions. The derivations are based on the representation of interior and exterior sound fields using the simple source approach and determination of the simple source solutions with uniform rotation. Numerical simulations of rotating sound fields are presented to verify the theory.
Active room compensation for sound reinforcement using sound field separation techniques.
Heuchel, Franz M; Fernandez-Grande, Efren; Agerkvist, Finn T; Shabalina, Elena
2018-03-01
This work investigates how the sound field created by a sound reinforcement system can be controlled at low frequencies. An indoor control method is proposed which actively absorbs the sound incident on a reflecting boundary using an array of secondary sources. The sound field is separated into incident and reflected components by a microphone array close to the secondary sources, enabling the minimization of reflected components by means of optimal signals for the secondary sources. The method is purely feed-forward and assumes constant room conditions. Three different sound field separation techniques for the modeling of the reflections are investigated based on plane wave decomposition, equivalent sources, and the Spatial Fourier transform. Simulations and an experimental validation are presented, showing that the control method performs similarly well at enhancing low frequency responses with the three sound separation techniques. Resonances in the entire room are reduced, although the microphone array and secondary sources are confined to a small region close to the reflecting wall. Unlike previous control methods based on the creation of a plane wave sound field, the investigated method works in arbitrary room geometries and primary source positions.
NASA Astrophysics Data System (ADS)
Chen, Xiaol; Guo, Bei; Tuo, Jinliang; Zhou, Ruixin; Lu, Yang
2017-08-01
Nowadays, people are paying more and more attention to the noise reduction of household refrigerator compressor. This paper established a sound field bounded by compressor shell and ISO3744 standard field points. The Acoustic Transfer Vector (ATV) in the sound field radiated by a refrigerator compressor shell were calculated which fits the test result preferably. Then the compressor shell surface is divided into several parts. Based on Acoustic Transfer Vector approach, the sound pressure contribution to the field points and the sound power contribution to the sound field of each part were calculated. To obtain the noise radiation in the sound field, the sound pressure cloud charts were analyzed, and the contribution curves in different frequency of each part were acquired. Meanwhile, the sound power contribution of each part in different frequency was analyzed, to ensure those parts where contributes larger sound power. Through the analysis of acoustic contribution, those parts where radiate larger noise on the compressor shell were determined. This paper provides a credible and effective approach on the structure optimal design of refrigerator compressor shell, which is meaningful in the noise and vibration reduction.
A quasi two-dimensional model for sound attenuation by the sonic crystals.
Gupta, A; Lim, K M; Chew, C H
2012-10-01
Sound propagation in the sonic crystal (SC) along the symmetry direction is modeled by sound propagation through a variable cross-sectional area waveguide. A one-dimensional (1D) model based on the Webster horn equation is used to obtain sound attenuation through the SC. This model is compared with two-dimensional (2D) finite element simulation and experiment. The 1D model prediction of frequency band for sound attenuation is found to be shifted by around 500 Hz with respect to the finite element simulation. The reason for this shift is due to the assumption involved in the 1D model. A quasi 2D model is developed for sound propagation through the waveguide. Sound pressure profiles from the quasi 2D model are compared with the finite element simulation and the 1D model. The result shows significant improvement over the 1D model and is in good agreement with the 2D finite element simulation. Finally, sound attenuation through the SC is computed based on the quasi 2D model and is found to be in good agreement with the finite element simulation. The quasi 2D model provides an improved method to calculate sound attenuation through the SC.
The Role of Soundscape in Nature-Based Rehabilitation: A Patient Perspective.
Cerwén, Gunnar; Pedersen, Eja; Pálsdóttir, Anna-María
2016-12-11
Nature-based rehabilitation (NBR) has convincing support in research, yet the underlying mechanisms are not fully understood. The present study sought to increase understanding of the role of soundscapes in NBR, an aspect paid little attention thus far. Transcribed interviews with 59 patients suffering from stress-related mental disorders and undergoing a 12-week therapy programme in the rehabilitation garden in Alnarp, Sweden, were analysed using Interpretative Phenomenology Analysis (IPA). Described sounds were categorised as natural, technological or human. The results showed that patients frequently referred to natural sounds as being part of a pleasant and "quiet" experience that supported recovery and induced "soft fascination". Technological sounds were experienced as disturbing, while perception of human sounds varied depending on loudness and the social context. The study further uncovered how sound influenced patients' behaviour and experiences in the garden, through examination of three cross-theme dimensions that materialised in the study; sound in relation to overall perception, sound in relation to garden usage, and increased susceptibility to sound. The findings are discussed in relation to NBR; the need for a more nuanced understanding of susceptibility to sound among people suffering from mental fatigue was identified and design considerations for future rehabilitation gardens were formulated.
Airborne sound transmission loss characteristics of woodframe construction
Fred F. Rudder
1985-01-01
This report summarizes the available data on the airborne sound transmission loss properties of wood-frame construction and evaluates the methods for predicting the airborne sound transmission loss. The first part of the report comprises a summary of sound transmission loss data for wood-frame interior walls and floor-ceiling construction. Data bases describing the...
Global patterns of changes in underwater sound transmission caused by ocean acidification
NASA Astrophysics Data System (ADS)
Ilyina, T.; Zeebe, R. E.; Brewer, P. G.
2009-04-01
Oceanic uptake of man-made CO2 leads to a decrease in the ocean pH and carbonate saturation state. This processes, known as ocean acidification is expected to have adverse effects on a variety of marine organisms. A surprising consequence of ocean acidification, which has gone widely unrecognized, is its effect on underwater sound transmission. Low-frequency sound absorption in the ocean occurs due to chemical relaxation of the pH-dependent boric acid-borate ion reaction. As ocean pH drops, sound absorption in the audible range decreases. The decreased sound absorption will amplify ambient noise levels, and enhance long distance sound transmission, although its exact environmental impact is uncertain. Changes in the underwater sound absorption will affect the operation of scientific, commercial, and naval applications that are based on ocean acoustics, with yet unknown consequences for marine life. We project these changes using a global biogeochemical model (HAMOCC), which is forced by the anthropogenic CO2 emissions during the years 1800-2300. Based on model projections, we quantify when and where in the ocean these ocean chemistry induced perturbations in sound absorption will occur.
Research and Implementation of Heart Sound Denoising
NASA Astrophysics Data System (ADS)
Liu, Feng; Wang, Yutai; Wang, Yanxiang
Heart sound is one of the most important signals. However, the process of getting heart sound signal can be interfered with many factors outside. Heart sound is weak electric signal and even weak external noise may lead to the misjudgment of pathological and physiological information in this signal, thus causing the misjudgment of disease diagnosis. As a result, it is a key to remove the noise which is mixed with heart sound. In this paper, a more systematic research and analysis which is involved in heart sound denoising based on matlab has been made. The study of heart sound denoising based on matlab firstly use the powerful image processing function of matlab to transform heart sound signals with noise into the wavelet domain through wavelet transform and decomposition these signals in muli-level. Then for the detail coefficient, soft thresholding is made using wavelet transform thresholding to eliminate noise, so that a signal denoising is significantly improved. The reconstructed signals are gained with stepwise coefficient reconstruction for the processed detail coefficient. Lastly, 50HZ power frequency and 35 Hz mechanical and electrical interference signals are eliminated using a notch filter.
NASA Astrophysics Data System (ADS)
Yun, Dong-Un; Lee, Sang-Kwon
2017-06-01
In this paper, we present a novel method for an objective evaluation of knocking noise emitted by diesel engines based on the temporal and frequency masking theory. The knocking sound of a diesel engine is a vibro-acoustic sound correlated with the high-frequency resonances of the engine structure and a periodic impulsive sound with amplitude modulation. Its period is related to the engine speed and includes specific frequency bands related to the resonances of the engine structure. A knocking sound with the characteristics of a high-frequency impulsive wave can be masked by low-frequency sounds correlated with the harmonics of the firing frequency and broadband noise. The degree of modulation of the knocking sound signal was used for such objective evaluations in previous studies, without considering the masking effect. However, the frequency masking effect must be considered for the objective evaluation of the knocking sound. In addition to the frequency masking effect, the temporal masking effect occurs because the period of the knocking sound changes according to the engine speed. Therefore, an evaluation method considering the temporal and frequency masking effect is required to analyze the knocking sound objectively. In this study, an objective evaluation method considering the masking effect was developed based on the masking theory of sound and signal processing techniques. The method was applied successfully for the objective evaluation of the knocking sound of a diesel engine.
NASA Astrophysics Data System (ADS)
Zhang, Dashan; Guo, Jie; Jin, Yi; Zhu, Chang'an
2017-09-01
High-speed cameras provide full field measurement of structure motions and have been applied in nondestructive testing and noncontact structure monitoring. Recently, a phase-based method has been proposed to extract sound-induced vibrations from phase variations in videos, and this method provides insights into the study of remote sound surveillance and material analysis. An efficient singular value decomposition (SVD)-based approach is introduced to detect sound-induced subtle motions from pixel intensities in silent high-speed videos. A high-speed camera is initially applied to capture a video of the vibrating objects stimulated by sound fluctuations. Then, subimages collected from a small region on the captured video are reshaped into vectors and reconstructed to form a matrix. Orthonormal image bases (OIBs) are obtained from the SVD of the matrix; available vibration signal can then be obtained by projecting subsequent subimages onto specific OIBs. A simulation test is initiated to validate the effectiveness and efficiency of the proposed method. Two experiments are conducted to demonstrate the potential applications in sound recovery and material analysis. Results show that the proposed method efficiently detects subtle motions from the video.
A Numerical Experiment on the Role of Surface Shear Stress in the Generation of Sound
NASA Technical Reports Server (NTRS)
Shariff, Karim; Wang, Meng; Merriam, Marshal (Technical Monitor)
1996-01-01
The sound generated due to a localized flow over an infinite flat surface is considered. It is known that the unsteady surface pressure, while appearing in a formal solution to the Lighthill equation, does not constitute a source of sound but rather represents the effect of image quadrupoles. The question of whether a similar surface shear stress term constitutes a true source of dipole sound is less settled. Some have boldly assumed it is a true source while others have argued that, like the surface pressure, it depends on the sound field (via an acoustic boundary layer) and is therefore not a true source. A numerical experiment based on the viscous, compressible Navier-Stokes equations was undertaken to investigate the issue. A small region of a wall was oscillated tangentially. The directly computed sound field was found to to agree with an acoustic analogy based calculation which regards the surface shear as an acoustically compact dipole source of sound.
NASA Astrophysics Data System (ADS)
Kella, Vara Prasad; Ghosh, Joydeep; Chattopadhyay, Prabal; Sharma, Devendra; Saxena, Yogesh
2017-10-01
Recent experimental measurements of ion flow speeds near the sheath edge of two-ion species plasma shows that, the ions reach the sheath edge with common sound speed other than their individual Bohm speeds at nearly equal ion concentrations. Baalrud et al., explain these results on the basis of ion-ion two-stream instability enhanced collisional friction between the ions. Some authors stipulate the existence of the instability indirectly, by measuring the ion flow speeds near the sheath edge. In these experiments, the instability is directly observed from the floating potential fluctuations from Langmuir probe placed near the sheath edge and from grid in Ar +He plasma. The frequency spectra shows broad band peaks with central frequency in the range 150-200 kHz. The intensity of the instability maximizes in the plasma produced with approximately equal ion concentrations of both the ion species. The frequency and amplitude of the peak decreases as the He+ to Ar+ concentration ratio decreases from unity. The phase velocity of the wave is measured as 11 +/-2 km/s and identified to be twice the ion-sound speed in the bulk ( 6.3 km/s), which is good agreement with earlier results of IAWs. The measured wave number and frequencies are compared with the theoretical dispersion relations. These observations confirm the existence of ion-ion co-stream instability in sheath-presheath of two-ion species plasma.
Underwater Sound Propagation from Marine Pile Driving.
Reyff, James A
2016-01-01
Pile driving occurs in a variety of nearshore environments that typically have very shallow-water depths. The propagation of pile-driving sound in water is complex, where sound is directly radiated from the pile as well as through the ground substrate. Piles driven in the ground near water bodies can produce considerable underwater sound energy. This paper presents examples of sound propagation through shallow-water environments. Some of these examples illustrate the substantial variation in sound amplitude over time that can be critical to understand when computing an acoustic-based safety zone for aquatic species.
Kastelein, Ronald A; van der Heul, Sander; Verboom, Willem C; Triesscheijn, Rob J V; Jennings, Nancy V
2006-02-01
To prevent grounding of ships and collisions between ships in shallow coastal waters, an underwater data collection and communication network (ACME) using underwater sounds to encode and transmit data is currently under development. Marine mammals might be affected by ACME sounds since they may use sound of a similar frequency (around 12 kHz) for communication, orientation, and prey location. If marine mammals tend to avoid the vicinity of the acoustic transmitters, they may be kept away from ecologically important areas by ACME sounds. One marine mammal species that may be affected in the North Sea is the harbour seal (Phoca vitulina). No information is available on the effects of ACME-like sounds on harbour seals, so this study was carried out as part of an environmental impact assessment program. Nine captive harbour seals were subjected to four sound types, three of which may be used in the underwater acoustic data communication network. The effect of each sound was judged by comparing the animals' location in a pool during test periods to that during baseline periods, during which no sound was produced. Each of the four sounds could be made into a deterrent by increasing its amplitude. The seals reacted by swimming away from the sound source. The sound pressure level (SPL) at the acoustic discomfort threshold was established for each of the four sounds. The acoustic discomfort threshold is defined as the boundary between the areas that the animals generally occupied during the transmission of the sounds and the areas that they generally did not enter during transmission. The SPLs at the acoustic discomfort thresholds were similar for each of the sounds (107 dB re 1 microPa). Based on this discomfort threshold SPL, discomfort zones at sea for several source levels (130-180 dB re 1 microPa) of the sounds were calculated, using a guideline sound propagation model for shallow water. The discomfort zone is defined as the area around a sound source that harbour seals are expected to avoid. The definition of the discomfort zone is based on behavioural discomfort, and does not necessarily coincide with the physical discomfort zone. Based on these results, source levels can be selected that have an acceptable effect on harbour seals in particular areas. The discomfort zone of a communication sound depends on the sound, the source level, and the propagation characteristics of the area in which the sound system is operational. The source level of the communication system should be adapted to each area (taking into account the width of a sea arm, the local sound propagation, and the importance of an area to the affected species). The discomfort zone should not coincide with ecologically important areas (for instance resting, breeding, suckling, and feeding areas), or routes between these areas.
The Problems with "Noise Numbers" for Wind Farm Noise Assessment
ERIC Educational Resources Information Center
Thorne, Bob
2011-01-01
Human perception responds primarily to sound character rather than sound level. Wind farms are unique sound sources and exhibit special audible and inaudible characteristics that can be described as modulating sound or as a tonal complex. Wind farm compliance measures based on a specified noise number alone will fail to address problems with noise…
2004-2006 Puget Sound Traffic Choices Study | Transportation Secure Data
Center | NREL 04-2006 Puget Sound Traffic Choices Study 2004-2006 Puget Sound Traffic Choices Study The 2004-2006 Puget Sound Traffic Choices Study tested the hypothesis that time-of-day variable Administration for a pilot project on congestion-based tolling. Methodology To test the hypothesis, the study
Code of Federal Regulations, 2010 CFR
2010-07-01
... Gulf of Mexico adjacent to Santa Rosa Island, Air Force Proving Ground Command, Eglin Air Force Base... Sound and Gulf of Mexico adjacent to Santa Rosa Island, Air Force Proving Ground Command, Eglin Air Force Base, Fla. (a) The danger zones—(1) Prohibited area. Waters of Santa Rosa Sound and Gulf of Mexico...
Acoustic Tomography of the Atmospheric Surface Layer
2014-11-28
Report Title Acoustic tomography of the atmospheric surface layer (ASL) is based on the measurements of the travel times of sound propagation between...SECURITY CLASSIFICATION OF: Acoustic tomography of the atmospheric surface layer (ASL) is based on the measurements of the travel times of sound ...organ. In the case of acoustic tomography of the atmospheric surface layer (ASL), the travel times of sound propagation between speakers and
2001-10-25
wavelet decomposition of signals and classification using neural network. Inputs to the system are the heart sound signals acquired by a stethoscope in a...Proceedings. pp. 415–418, 1990. [3] G. Ergun, “An intelligent diagnostic system for interpretation of arterpartum fetal heart rate tracings based on ANNs and...AN INTELLIGENT PATTERN RECOGNITION SYSTEM BASED ON NEURAL NETWORK AND WAVELET DECOMPOSITION FOR INTERPRETATION OF HEART SOUNDS I. TURKOGLU1, A
Sound reduction by metamaterial-based acoustic enclosure
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yao, Shanshan; Li, Pei; Zhou, Xiaoming
In many practical systems, acoustic radiation control on noise sources contained within a finite volume by an acoustic enclosure is of great importance, but difficult to be accomplished at low frequencies due to the enhanced acoustic-structure interaction. In this work, we propose to use acoustic metamaterials as the enclosure to efficiently reduce sound radiation at their negative-mass frequencies. Based on a circularly-shaped metamaterial model, sound radiation properties by either central or eccentric sources are analyzed by numerical simulations for structured metamaterials. The parametric analyses demonstrate that the barrier thickness, the cavity size, the source type, and the eccentricity of themore » source have a profound effect on the sound reduction. It is found that increasing the thickness of the metamaterial barrier is an efficient approach to achieve large sound reduction over the negative-mass frequencies. These results are helpful in designing highly efficient acoustic enclosures for blockage of sound in low frequencies.« less
Personal sleep pattern visualization using sequence-based kernel self-organizing map on sound data.
Wu, Hongle; Kato, Takafumi; Yamada, Tomomi; Numao, Masayuki; Fukui, Ken-Ichi
2017-07-01
We propose a method to discover sleep patterns via clustering of sound events recorded during sleep. The proposed method extends the conventional self-organizing map algorithm by kernelization and sequence-based technologies to obtain a fine-grained map that visualizes the distribution and changes of sleep-related events. We introduced features widely applied in sound processing and popular kernel functions to the proposed method to evaluate and compare performance. The proposed method provides a new aspect of sleep monitoring because the results demonstrate that sound events can be directly correlated to an individual's sleep patterns. In addition, by visualizing the transition of cluster dynamics, sleep-related sound events were found to relate to the various stages of sleep. Therefore, these results empirically warrant future study into the assessment of personal sleep quality using sound data. Copyright © 2017 Elsevier B.V. All rights reserved.
Caldwell, William Scott
2001-01-01
Data collected at three sites in Currituck Sound and three tributary sites between March 1, 1998, and February 28, 1999, were used to describe hydrologic and salinity characteristics of Currituck Sound. Water levels and salinity were measured at West Neck Creek at Pungo and at Albemarle and Chesapeake Canal near Princess Anne in Virginia, and at Coinjock, Bell Island, Poplar Branch, and Point Harbor in North Carolina. Flow velocity also was measured at the West Neck Creek and Coinjock sites.The maximum water-level range during the study period was observed near the lower midpoint of Currituck Sound at Poplar Branch. Generally, water levels at all sites were highest during March and April, and lowest during November and December. Winds from the south typically produced higher water levels in Currituck Sound, whereas winds from the north typically produced lower water levels. Although wind over Currituck Sound is associated with fluctuations in water level within the sound, other mechanisms, such as the effects of wind on Albemarle Sound and on other water bodies south of Currituck Sound, likely affect low-frequency water-level variations in Currituck Sound.Flow in West Neck Creek ranged from 313 cubic feet per second to the south to -227 cubic feet per second to the north (negative indicates flow to the north). Flow at the Coinjock site ranged from 15,300 cubic feet per second to the south to -11,700 cubic feet per second to the north. Flow was to the south 68 percent of the time at the West Neck Creek site and 44 percent of the time at the Coinjock site. Daily flow volumes were calculated as the sum of the instantaneous flow volumes. The West Neck Creek site had a cumulative flow volume to the south of 7.69 x 108 cubic feet for the period March 1, 1998, to February 28, 1999; the Coinjock site had a cumulative flow volume to the north of -1.33 x 1010 cubic feet for the same study period.Wind direction and speed influence flow at the West Neck Creek and Coinjock sites, whereas precipitation alone has little effect on flow at these sites. Flow at the West Neck Creek site is semidiurnal but is affected by wind direction and speed. Flow to the south (positive flow) was associated with wind speeds averaging more than 15 miles per hour from the northwest; flow to the north (negative flow) was associated with wind speeds averaging more than 15 miles per hour from the south and southwest. Flow at the Coinjock site reacted in a more unpredictable manner and was not affected by winds or tides in the same manner as West Neck Creek, with few tidal characteristics evident in the record.Throughout the study period, maximum salinity exceeded 3.5 parts per thousand at all sites; however, mean and median salinities were below 3.5 parts per thousand at all sites except the Point Harbor site (3.6 and 4.2 parts per thousand, respectively) at the southern end of the sound. Salinities were less than or equal to 3.5 parts per thousand nearly 100 percent of the time at the Bell Island and Poplar Branch sites in Currituck Sound and about 86 percent of the time at the Albemarle and Chesapeake Canal site north of the sound. Salinity at the West Neck Creek and Coinjock sites was less than or equal to 3.5 parts per thousand about 82 percent of the time.During this study, prevailing winds from the north were associated with flow to the south and tended to increase salinity at the West Neck Creek and the Albemarle and Chesapeake Canal sites. Conversely, these same winds tended to decrease salinity at the other sites. Prevailing winds from the south and southwest were associated with flow to the north and tended to increase salinity at the Poplar Branch and Point Harbor sites in Currituck Sound and at the Coinjock site, but these same winds tended to decrease salinity at the West Neck Creek and the Albemarle and Chesapeake Canal sites. The greatest variations in salinity were observed at the northernmost site, West Neck Creek, and thesouthernmost site, Point Harbor. The least variation in salinity was observed at the upper midpoint of the sound at the Bell Island site.Daily salt loads were computed for 364 days at the West Neck Creek site and 348 days at the Coinjock site from March 1, 1998, to February 28, 1999. The cumulative salt load at West Neck Creek was 28,170 tons to the south, and the cumulative salt load at the Coinjock site was -872,750 tons to the north.The cumulative salt load passing the West Neck Creek site during the study period would be 0.01 part per thousand if uniformly distributed throughout the sound (approximately 489,600 acre-feet in North Carolina). If the cumulative salt load passing the Coinjock site were uniformly distributed throughout the sound, the salinity in the sound would be 0.32 part per thousand. The net transport at the West Neck Creek and Coinjock sites indicates inflow of salt into the sound. A constant inflow of freshwater from tributaries and ground-water sources also occurs; however, the net flow volumes from these freshwater sources are not documented, and the significance of these freshwater inflows toward diluting the net import of salt into the sound is beyond the scope of this study.
Design and Implementation of Sound Searching Robots in Wireless Sensor Networks
Han, Lianfu; Shen, Zhengguang; Fu, Changfeng; Liu, Chao
2016-01-01
A sound target-searching robot system which includes a 4-channel microphone array for sound collection, magneto-resistive sensor for declination measurement, and a wireless sensor networks (WSN) for exchanging information is described. It has an embedded sound signal enhancement, recognition and location method, and a sound searching strategy based on a digital signal processor (DSP). As the wireless network nodes, three robots comprise the WSN a personal computer (PC) in order to search the three different sound targets in task-oriented collaboration. The improved spectral subtraction method is used for noise reduction. As the feature of audio signal, Mel-frequency cepstral coefficient (MFCC) is extracted. Based on the K-nearest neighbor classification method, we match the trained feature template to recognize sound signal type. This paper utilizes the improved generalized cross correlation method to estimate time delay of arrival (TDOA), and then employs spherical-interpolation for sound location according to the TDOA and the geometrical position of the microphone array. A new mapping has been proposed to direct the motor to search sound targets flexibly. As the sink node, the PC receives and displays the result processed in the WSN, and it also has the ultimate power to make decision on the received results in order to improve their accuracy. The experiment results show that the designed three-robot system implements sound target searching function without collisions and performs well. PMID:27657088
Design of Alarm Sound of Home Care Equipment Based on Age-related Auditory Sense
NASA Astrophysics Data System (ADS)
Shibano, Jun-Ichi; Tadano, Shigeru; Kaneko, Hirotaka
A wide variety of home care equipment has been developed to support the independent lifestyle and care taking of elderly persons. Almost all of the equipment has an alarm designed to alert a care person or to sound a warning in case of an emergency. Due to the fact that aging human beings' senses physiologically, weaken and deteriorate, each alarm's sound must be designed to account for the full range of elderly person's hearing loss. Since the alarms are usually heard indoors, it is also necessary to evaluate the relationship between the basic characteristics of the sounds and living area's layout. In this study, we investigated the sounds of various alarms of the home care equipment based on both the age-related hearing characteristics of elderly persons and the propagation property of the sounds indoors. As a result, it was determined that the hearing characteristics of elderly persons are attuned to sounds which have a frequency from 700Hz to 1kHz, and it was learned that the indoor absorption ratio of sound is smallest when the frequency is 1kHz. Therefore, a frequency of 1kHz is good for the alarm sound of home care equipment. A flow chart to design the alarm sound of home care equipment was proposed, taking into account the extent of age-related auditory sense deterioration.
Design and Implementation of Sound Searching Robots in Wireless Sensor Networks.
Han, Lianfu; Shen, Zhengguang; Fu, Changfeng; Liu, Chao
2016-09-21
A sound target-searching robot system which includes a 4-channel microphone array for sound collection, magneto-resistive sensor for declination measurement, and a wireless sensor networks (WSN) for exchanging information is described. It has an embedded sound signal enhancement, recognition and location method, and a sound searching strategy based on a digital signal processor (DSP). As the wireless network nodes, three robots comprise the WSN a personal computer (PC) in order to search the three different sound targets in task-oriented collaboration. The improved spectral subtraction method is used for noise reduction. As the feature of audio signal, Mel-frequency cepstral coefficient (MFCC) is extracted. Based on the K-nearest neighbor classification method, we match the trained feature template to recognize sound signal type. This paper utilizes the improved generalized cross correlation method to estimate time delay of arrival (TDOA), and then employs spherical-interpolation for sound location according to the TDOA and the geometrical position of the microphone array. A new mapping has been proposed to direct the motor to search sound targets flexibly. As the sink node, the PC receives and displays the result processed in the WSN, and it also has the ultimate power to make decision on the received results in order to improve their accuracy. The experiment results show that the designed three-robot system implements sound target searching function without collisions and performs well.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Koukoulas, Triantafillos, E-mail: triantafillos.koukoulas@npl.co.uk; Piper, Ben
Since the introduction of the International System of Units (the SI system) in 1960, weights, measures, standardised approaches, procedures, and protocols have been introduced, adapted, and extensively used. A major international effort and activity concentrate on the definition and traceability of the seven base SI units in terms of fundamental constants, and consequently those units that are derived from the base units. In airborne acoustical metrology and for the audible range of frequencies up to 20 kHz, the SI unit of sound pressure, the pascal, is realised indirectly and without any knowledge or measurement of the sound field. Though themore » principle of reciprocity was originally formulated by Lord Rayleigh nearly two centuries ago, it was devised in the 1940s and eventually became a calibration standard in the 1960s; however, it can only accommodate a limited number of acoustic sensors of specific types and dimensions. International standards determine the device sensitivity either through coupler or through free-field reciprocity but rely on the continuous availability of specific acoustical artefacts. Here, we show an optical method based on gated photon correlation spectroscopy that can measure sound pressures directly and absolutely in fully anechoic conditions, remotely, and without disturbing the propagating sound field. It neither relies on the availability or performance of any measurement artefact nor makes any assumptions of the device geometry and sound field characteristics. Most importantly, the required units of sound pressure and microphone sensitivity may now be experimentally realised, thus providing direct traceability to SI base units.« less
Lee, Kyoung-Ryul; Jang, Sung Hwan; Jung, Inhwa
2018-08-10
We investigated the acoustic performance of electrostatic sound-generating devices consisting of bi-layer graphene on polyimide film. The total sound pressure level (SPL) of the sound generated from the devices was measured as a function of source frequency by sweeping, and frequency spectra were measured at 1/3 octave band frequencies. The relationship between various operation conditions and total SPL was determined. In addition, the effects of changing voltage level, adding a DC offset, and using two pairs of electrodes were evaluated. It should be noted that two pairs of electrode operations improved sound generation by about 10 dB over all frequency ranges compared with conventional operation. As for the sound-generating capability, total SPL was 70 dBA at 4 kHz when an AC voltage of 100 V pp was applied with a DC offset of 100 V. Acoustic characteristics differed from other types of graphene-based sound generators, such as graphene thermoacoustic devices and graphene polyvinylidene fluoride devices. The effects of diameter and distance between electrodes were also studied, and we found that diameter greatly influenced the frequency response. We anticipate that the design information provided in this paper, in addition to describing key parameters of electrostatic sound-generating devices, will facilitate the commercial development of electrostatic sound-generating systems.
The Role of Soundscape in Nature-Based Rehabilitation: A Patient Perspective
Cerwén, Gunnar; Pedersen, Eja; Pálsdóttir, Anna María
2016-01-01
Nature-based rehabilitation (NBR) has convincing support in research, yet the underlying mechanisms are not fully understood. The present study sought to increase understanding of the role of soundscapes in NBR, an aspect paid little attention thus far. Transcribed interviews with 59 patients suffering from stress-related mental disorders and undergoing a 12-week therapy programme in the rehabilitation garden in Alnarp, Sweden, were analysed using Interpretative Phenomenology Analysis (IPA). Described sounds were categorised as natural, technological or human. The results showed that patients frequently referred to natural sounds as being part of a pleasant and “quiet” experience that supported recovery and induced “soft fascination”. Technological sounds were experienced as disturbing, while perception of human sounds varied depending on loudness and the social context. The study further uncovered how sound influenced patients’ behaviour and experiences in the garden, through examination of three cross-theme dimensions that materialised in the study; sound in relation to overall perception, sound in relation to garden usage, and increased susceptibility to sound. The findings are discussed in relation to NBR; the need for a more nuanced understanding of susceptibility to sound among people suffering from mental fatigue was identified and design considerations for future rehabilitation gardens were formulated. PMID:27973437
Bai, Mingsian R; Li, Yi; Chiang, Yi-Hao
2017-10-01
A unified framework is proposed for analysis and synthesis of two-dimensional spatial sound field in reverberant environments. In the sound field analysis (SFA) phase, an unbaffled 24-element circular microphone array is utilized to encode the sound field based on the plane-wave decomposition. Depending on the sparsity of the sound sources, the SFA stage can be implemented in two manners. For sparse-source scenarios, a one-stage algorithm based on compressive sensing algorithm is utilized. Alternatively, a two-stage algorithm can be used, where the minimum power distortionless response beamformer is used to localize the sources and Tikhonov regularization algorithm is used to extract the source amplitudes. In the sound field synthesis (SFS), a 32-element rectangular loudspeaker array is employed to decode the target sound field using pressure matching technique. To establish the room response model, as required in the pressure matching step of the SFS phase, an SFA technique for nonsparse-source scenarios is utilized. Choice of regularization parameters is vital to the reproduced sound field. In the SFS phase, three SFS approaches are compared in terms of localization performance and voice reproduction quality. Experimental results obtained in a reverberant room are presented and reveal that an accurate room response model is vital to immersive rendering of the reproduced sound field.
Intelligent Systems Approaches to Product Sound Quality Analysis
NASA Astrophysics Data System (ADS)
Pietila, Glenn M.
As a product market becomes more competitive, consumers become more discriminating in the way in which they differentiate between engineered products. The consumer often makes a purchasing decision based on the sound emitted from the product during operation by using the sound to judge quality or annoyance. Therefore, in recent years, many sound quality analysis tools have been developed to evaluate the consumer preference as it relates to a product sound and to quantify this preference based on objective measurements. This understanding can be used to direct a product design process in order to help differentiate the product from competitive products or to establish an impression on consumers regarding a product's quality or robustness. The sound quality process is typically a statistical tool that is used to model subjective preference, or merit score, based on objective measurements, or metrics. In this way, new product developments can be evaluated in an objective manner without the laborious process of gathering a sample population of consumers for subjective studies each time. The most common model used today is the Multiple Linear Regression (MLR), although recently non-linear Artificial Neural Network (ANN) approaches are gaining popularity. This dissertation will review publicly available published literature and present additional intelligent systems approaches that can be used to improve on the current sound quality process. The focus of this work is to address shortcomings in the current paired comparison approach to sound quality analysis. This research will propose a framework for an adaptive jury analysis approach as an alternative to the current Bradley-Terry model. The adaptive jury framework uses statistical hypothesis testing to focus on sound pairings that are most interesting and is expected to address some of the restrictions required by the Bradley-Terry model. It will also provide a more amicable framework for an intelligent systems approach. Next, an unsupervised jury clustering algorithm is used to identify and classify subgroups within a jury who have conflicting preferences. In addition, a nested Artificial Neural Network (ANN) architecture is developed to predict subjective preference based on objective sound quality metrics, in the presence of non-linear preferences. Finally, statistical decomposition and correlation algorithms are reviewed that can help an analyst establish a clear understanding of the variability of the product sounds used as inputs into the jury study and to identify correlations between preference scores and sound quality metrics in the presence of non-linearities.
Algorithm Development for a Real-Time Military Noise Monitor
2006-03-24
Duration ESLM Enhanced Sound Level Meter ERDC-CERL Engineer Research and Development Center/Construction Engineering Research Laboratory FFT...Fast Fourier Transform FTIG Fort Indiantown Gap Kurt Kurtosis LD Larson Davis Leq Equivalent Sound Level L8eq 8-hr Equivalent...Sound Level Lpk Peak Sound Level m Spectral Slope MCBCL Marine Corps Base Camp Lejeune Neg Number of negative samples NI National
The detection of differences in the cues to distance by elderly hearing-impaired listeners
Akeroyd, Michael A.; Blaschke, Julia; Gatehouse, Stuart
2013-01-01
This experiment measured the capability of hearing-impaired individuals to discriminate differences in the cues to the distance of spoken sentences. The stimuli were generated synthetically, using a room-image procedure to calculate the direct sound and first 74 reflections for a source placed in a 7 × 9 m room, and then presenting each of those sounds individually through a circular array of 24 loudspeakers. Seventy-seven listeners participated, aged 22-83 years and with hearing levels from −5 to 59 dB HL. In conditions where a substantial change in overall level due to the inverse-square law was available as a cue, the elderly-hearing-impaired listeners did not perform any different from control groups. In other conditions where that cue was unavailable (so leaving the direct-to-reverberant relationship as a cue), either because the reverberant field dominated the direct sound or because the overall level had been artificially equalized, hearing-impaired listeners performed worse than controls. There were significant correlations with listeners’ self-reported distance capabilities as measured by the “SSQ” questionnaire [S. Gatehouse and W. Noble, Int. J. Audiol. 43, 85-99 (2004)]. The results demonstrate that hearing-impaired listeners show deficits in the ability to use some of the cues which signal auditory distance. PMID:17348530
Ultrasound visual feedback treatment and practice variability for residual speech sound errors
Preston, Jonathan L.; McCabe, Patricia; Rivera-Campos, Ahmed; Whittle, Jessica L.; Landry, Erik; Maas, Edwin
2014-01-01
Purpose The goals were to (1) test the efficacy of a motor-learning based treatment that includes ultrasound visual feedback for individuals with residual speech sound errors, and (2) explore whether the addition of prosodic cueing facilitates speech sound learning. Method A multiple baseline single subject design was used, replicated across 8 participants. For each participant, one sound context was treated with ultrasound plus prosodic cueing for 7 sessions, and another sound context was treated with ultrasound but without prosodic cueing for 7 sessions. Sessions included ultrasound visual feedback as well as non-ultrasound treatment. Word-level probes assessing untreated words were used to evaluate retention and generalization. Results For most participants, increases in accuracy of target sound contexts at the word level were observed with the treatment program regardless of whether prosodic cueing was included. Generalization between onset singletons and clusters was observed, as well as generalization to sentence-level accuracy. There was evidence of retention during post-treatment probes, including at a two-month follow-up. Conclusions A motor-based treatment program that includes ultrasound visual feedback can facilitate learning of speech sounds in individuals with residual speech sound errors. PMID:25087938
2015-09-01
OAT) and laser-induced ultrasound tomography (LUT) to obtain coregistered maps of tissue optical absorption and speed of sound , displayed within the...computed tomography (UST) can provide high-resolution anatomical images of breast lesions based on three complementary acoustic properties (speed-of- sound ...tomography (UST) can provide high-resolution anatomical images of breast lesions based on three complementary acoustic properties (speed-of- sound
NASA Astrophysics Data System (ADS)
Kulichkov, S. N.; Popov, O. Ye.; Mishenin, A. A.; Chunchuzov, I. P.; Chkhetiani, O. G.; Tsybulskaya, N. D.
2017-11-01
The atmospheric effect on the characteristics of infrasonic signals from explosions has been studied. New methods have been proposed to remotely estimate the energy of explosions using the data of infrasonic wave registration. One method is based on the law of conservation of acoustic pulse I, which is equal to the product of the wave profile area S/2 of the studied infrasonic signal and the distance to the source E I [kt] = 1.38 × 10-10 (I [kg/s])1.482. The second method is based on the relationship between the explosion energy and the dominant period T of the recorded signal, ET [kt] =1.02 × ( T [s]2/σ)3/2, where σ is a dimensionless distance used for determining the degree of manifestation of nonlinear effects in the propagation of sound along ray trajectories. When compared to the conventional E W (Whitaker's) relation, the advantage of the EI relation is that it can be used for pulsed sources located at an arbitrary height over the land surface and having an arbitrary form of the initial-pulse profile and for any type of infrasonic arrivals. A distinctive feature of the expression for E T is that the atmospheric effect on the characteristics of recorded infrasonic signals is explicitly taken into account. These methods have been tested using infrasonic data recorded at a distance of 322 km from the sources (30 explosions caused by a fire that occurred at the Pugachevo armory in Udmurtia on June 2, 2011). For the same explosion, empirical relations have been found between energy values obtained by different methods: E I = 1.107 × E W , E T = 2.201 × E I .
Application of subharmonics for active sound design of electric vehicles.
Gwak, Doo Young; Yoon, Kiseop; Seong, Yeolwan; Lee, Soogab
2014-12-01
The powertrain of electric vehicles generates an unfamiliar acoustical environment for customers. This paper seeks optimal interior sound for electric vehicles based on psychoacoustic knowledge and musical harmonic theory. The concept of inserting a virtual sound, which consists of the subharmonics of an existing high-frequency component, is suggested to improve sound quality. Subjective evaluation results indicate that the impression of interior sound can be enhanced in this manner. Increased appeal is achieved through two designed stimuli, which proves the effectiveness of the method proposed.
NASA Astrophysics Data System (ADS)
Eshach, Haim
2014-06-01
This article describes the development and field test of the Sound Concept Inventory Instrument (SCII), designed to measure middle school students' concepts of sound. The instrument was designed based on known students' difficulties in understanding sound and the history of science related to sound and focuses on two main aspects of sound: sound has material properties, and sound has process properties. The final SCII consists of 71 statements that respondents rate as either true or false and also indicate their confidence on a five-point scale. Administration to 355 middle school students resulted in a Cronbach alpha of 0.906, suggesting a high reliability. In addition, the average percentage of students' answers to statements that associate sound with material properties is significantly higher than the average percentage of statements associating sound with process properties (p <0.001). The SCII is a valid and reliable tool that can be used to determine students' conceptions of sound.
Model-based synthesis of aircraft noise to quantify human perception of sound quality and annoyance
NASA Astrophysics Data System (ADS)
Berckmans, D.; Janssens, K.; Van der Auweraer, H.; Sas, P.; Desmet, W.
2008-04-01
This paper presents a method to synthesize aircraft noise as perceived on the ground. The developed method gives designers the opportunity to make a quick and economic evaluation concerning sound quality of different design alternatives or improvements on existing aircraft. By presenting several synthesized sounds to a jury, it is possible to evaluate the quality of different aircraft sounds and to construct a sound that can serve as a target for future aircraft designs. The combination of using a sound synthesis method that can perform changes to a recorded aircraft sound together with executing jury tests allows to quantify the human perception of aircraft noise.
Light aircraft sound transmission studies - Noise reduction model
NASA Technical Reports Server (NTRS)
Atwal, Mahabir S.; Heitman, Karen E.; Crocker, Malcolm J.
1987-01-01
Experimental tests conducted on the fuselage of a single-engine Piper Cherokee light aircraft suggest that the cabin interior noise can be reduced by increasing the transmission loss of the dominant sound transmission paths and/or by increasing the cabin interior sound absorption. The validity of using a simple room equation model to predict the cabin interior sound-pressure level for different fuselage and exterior sound field conditions is also presented. The room equation model is based on the sound power flow balance for the cabin space and utilizes the measured transmitted sound intensity data. The room equation model predictions were considered good enough to be used for preliminary acoustical design studies.
Embedded System Implementation of Sound Localization in Proximal Region
NASA Astrophysics Data System (ADS)
Iwanaga, Nobuyuki; Matsumura, Tomoya; Yoshida, Akihiro; Kobayashi, Wataru; Onoye, Takao
A sound localization method in the proximal region is proposed, which is based on a low-cost 3D sound localization algorithm with the use of head-related transfer functions (HRTFs). The auditory parallax model is applied to the current algorithm so that more accurate HRTFs can be used for sound localization in the proximal region. In addition, head-shadowing effects based on rigid-sphere model are reproduced in the proximal region by means of a second-order IIR filter. A subjective listening test demonstrates the effectiveness of the proposed method. Embedded system implementation of the proposed method is also described claiming that the proposed method improves sound effects in the proximal region only with 5.1% increase of memory capacity and 8.3% of computational costs.
[Analysis of the heart sound with arrhythmia based on nonlinear chaos theory].
Ding, Xiaorong; Guo, Xingming; Zhong, Lisha; Xiao, Shouzhong
2012-10-01
In this paper, a new method based on the nonlinear chaos theory was proposed to study the arrhythmia with the combination of the correlation dimension and largest Lyapunov exponent, through computing and analyzing these two parameters of 30 cases normal heart sound and 30 cases with arrhythmia. The results showed that the two parameters of the heart sounds with arrhythmia were higher than those with the normal, and there was significant difference between these two kinds of heart sounds. That is probably due to the irregularity of the arrhythmia which causes the decrease of predictability, and it's more complex than the normal heart sound. Therefore, the correlation dimension and the largest Lyapunov exponent can be used to analyze the arrhythmia and for its feature extraction.
Controllability of a multichannel system
NASA Astrophysics Data System (ADS)
Ivanov, Sergei A.; Wang, Jun Min
2018-02-01
We consider the system consisting of K coupled acoustic channels with the different sound velocities cj. Channels are interacting at any point via the pressure and its time derivatives. Using the moment approach and the theory of exponential families with vector coefficients we establish two controllability results: the system is exactly controllable if (i) the control uj in the jth channel acts longer than the double travel time of a wave from the start to the end of the j-th channel; (ii) all controls uj act more than or equal to the maximal double travel time.
NASA Technical Reports Server (NTRS)
2004-01-01
This cutaway illustration shows the Saturn V S-IC (first) stage with detailed callouts of the components. The S-IC Stage is 138 feet long and 33 feet in diameter, producing 7,500,000 pounds of thrust through five F-1 engines that are powered by liquid oxygen and kerosene. Four of the engines are mounted on an outer ring and gimbal for control purposes. The fifth engine is rigidly mounted in the center. When ignited, the roar produced by the five engines equals the sound of 8,000,000 hi-fi sets.
2004-04-15
This cutaway illustration shows the Saturn V S-IC (first) stage with detailed callouts of the components. The S-IC Stage is 138 feet long and 33 feet in diameter, producing 7,500,000 pounds of thrust through five F-1 engines that are powered by liquid oxygen and kerosene. Four of the engines are mounted on an outer ring and gimbal for control purposes. The fifth engine is rigidly mounted in the center. When ignited, the roar produced by the five engines equals the sound of 8,000,000 hi-fi sets.
1967-01-01
This illustration shows a cutaway drawing with callouts of the major components for the S-IC (first) stage of the Saturn V launch vehicle. The S-IC stage is 138 feet long and 33 feet in diameter, producing more than 7,500,000 pounds of thrust through five F-1 engines powered by liquid oxygen and kerosene. Four of the engines are mounted on an outer ring and gimball for control purposes. The fifth engine is rigidly mounted in the center. When ignited, the roar produced by the five engines equals the sound of 8,000,000 hi-fi sets.
NASA Astrophysics Data System (ADS)
Ovsiannikov, Mikhail; Ovsiannikov, Sergei
2017-01-01
The paper presents the combined approach to noise mapping and visualizing of industrial facilities sound pollution using forward ray tracing method and thin-plate spline interpolation. It is suggested to cauterize industrial area in separate zones with similar sound levels. Equivalent local source is defined for range computation of sanitary zones based on ray tracing algorithm. Computation of sound pressure levels within clustered zones are based on two-dimension spline interpolation of measured data on perimeter and inside the zone.
Auditory displays as occasion setters.
Mckeown, Denis; Isherwood, Sarah; Conway, Gareth
2010-02-01
The aim of this study was to evaluate whether representational sounds that capture the richness of experience of a collision enhance performance in braking to avoid a collision relative to other forms of warnings in a driving simulator. There is increasing interest in auditory warnings that are informative about their referents. But as well as providing information about some intended object, warnings may be designed to set the occasion for a rich body of information about the outcomes of behavior in a particular context. These richly informative warnings may offer performance advantages, as they may be rapidly processed by users. An auditory occasion setter for a collision (a recording of screeching brakes indicating imminent collision) was compared with two other auditory warnings (an abstract and an "environmental" sound), a speech message, a visual display, and no warning in a fixed-base driving simulator as interfaces to a collision avoidance system. The main measure was braking response times at each of two headways (1.5 s and 3 s) to a lead vehicle. The occasion setter demonstrated statistically significantly faster braking responses at each headway in 8 out of 10 comparisons (with braking responses equally fast to the abstract warning at 1.5 s and the environmental warning at 3 s). Auditory displays that set the occasion for an outcome in a particular setting and for particular behaviors may offer small but critical performance enhancements in time-critical applications. The occasion setter could be applied in settings where speed of response by users is of the essence.
[An implantable microphone for electronic hearing aids].
Leysieffer, H; Müller, G; Zenner, H P
1997-10-01
Fully implantable hearing aids and cochlea implants of the future require an implantable microphone. A hermetically sealed implantable microphone based on the idea of a microphone implanted in the posterior wall of the auditory canal, as suggested by Ohno et al. in 1988, is presented. Through consistent technological and clinical design optimization, it was possible to achieve a membrane diameter of only 4.5 mm (as opposed to 8 mm in the Japanese system) and a significant volume reduction of nearly 50%. The microphone weights only 0.4 g. In spite of this miniaturization, the performance characteristics of the microphone equal those of the Japanese model or are superior. The sound-pressure transfer function shows a very small ripple and the bandwidth amounts to approximately 10 kHz. Because of its high tuning and high no-load resonance frequency, the microphone is mostly insensitive to post-operational changes to the loading mass on the microphone membrane initiated by the covering skin of the auditory canal. The sound-pressure transfer factor at 1000 Hz is approximately 1.5 mV/Pa. Using different manufacturing technologies, this value can be increased in the range of 6-8 dB with a corresponding reduction in bandwidth. Due to the small mass, the microphone is highly insensitive to environmental mechanical disturbances. The module is made of pure titanium and is hermetically sealed according to Mil-Std 883 D. Full metal encapsulation and additional internal electronic components protect the microphone well against environmental electromagnetic influences (EMC).
Interactive physically-based sound simulation
NASA Astrophysics Data System (ADS)
Raghuvanshi, Nikunj
The realization of interactive, immersive virtual worlds requires the ability to present a realistic audio experience that convincingly compliments their visual rendering. Physical simulation is a natural way to achieve such realism, enabling deeply immersive virtual worlds. However, physically-based sound simulation is very computationally expensive owing to the high-frequency, transient oscillations underlying audible sounds. The increasing computational power of desktop computers has served to reduce the gap between required and available computation, and it has become possible to bridge this gap further by using a combination of algorithmic improvements that exploit the physical, as well as perceptual properties of audible sounds. My thesis is a step in this direction. My dissertation concentrates on developing real-time techniques for both sub-problems of sound simulation: synthesis and propagation. Sound synthesis is concerned with generating the sounds produced by objects due to elastic surface vibrations upon interaction with the environment, such as collisions. I present novel techniques that exploit human auditory perception to simulate scenes with hundreds of sounding objects undergoing impact and rolling in real time. Sound propagation is the complementary problem of modeling the high-order scattering and diffraction of sound in an environment as it travels from source to listener. I discuss my work on a novel numerical acoustic simulator (ARD) that is hundred times faster and consumes ten times less memory than a high-accuracy finite-difference technique, allowing acoustic simulations on previously-intractable spaces, such as a cathedral, on a desktop computer. Lastly, I present my work on interactive sound propagation that leverages my ARD simulator to render the acoustics of arbitrary static scenes for multiple moving sources and listener in real time, while accounting for scene-dependent effects such as low-pass filtering and smooth attenuation behind obstructions, reverberation, scattering from complex geometry and sound focusing. This is enabled by a novel compact representation that takes a thousand times less memory than a direct scheme, thus reducing memory footprints to fit within available main memory. To the best of my knowledge, this is the only technique and system in existence to demonstrate auralization of physical wave-based effects in real-time on large, complex 3D scenes.
Effect of real-world sounds on protein crystallization.
Zhang, Chen-Yan; Liu, Yue; Tian, Xu-Hua; Liu, Wen-Jing; Li, Xiao-Yu; Yang, Li-Xue; Jiang, Han-Jun; Han, Chong; Chen, Ke-An; Yin, Da-Chuan
2018-06-01
Protein crystallization is sensitive to the environment, while audible sound, as a physical and environmental factor during the entire process, is always ignored. We have previously reported that protein crystallization can be affected by a computer-generated monotonous sound with fixed frequency and amplitude. However, real-world sounds are not so simple but are complicated by parameters (frequency, amplitude, timbre, etc.) that vary over time. In this work, from three sound categories (music, speech, and environmental sound), we selected 26 different sounds and evaluated their effects on protein crystallization. The correlation between the sound parameters and the crystallization success rate was studied mathematically. The results showed that the real-world sounds, similar to the artificial monotonous sounds, could not only affect protein crystallization, but also improve crystal quality. Crystallization was dependent not only on the frequency, amplitude, volume, irradiation time, and overall energy of the sounds but also on their spectral characteristics. Based on these results, we suggest that intentionally applying environmental sound may be a simple and useful tool to promote protein crystallization. Copyright © 2018. Published by Elsevier B.V.
At-sea distribution of Spectacled Eiders: A 120-year-old mystery resolved
Petersen, Margaret R.; Larned, William W.; Douglas, David C.
1999-01-01
The at-sea distribution of the threatened Spectacled Eider (Somateria fischeri) has remained largely undocumented. We identified migration corridors, staging and molting areas, and wintering areas of adult Spectacled Eiders using implanted satellite-transmitters in birds from each of the three extant breeding grounds (North Slope and Yukon-Kuskokwim Delta in Alaska and arctic Russia). Based on transmitter locations, we conducted aerial surveys to provide visual confirmation of eider flocks and to estimate numbers of birds. We identified two principal molting and staging areas off coastal Alaska (Ledyard Bay and eastern Norton Sound) and two off coastal Russia (Mechigmenskiy Bay on the eastern Chukotka Peninsula, and the area between the Indigirka and Kolyma deltas in the Republic of Sakha). We estimated that >10,000 birds molt and stage in monospecific flocks at Mechigmenskiy and Ledyard bays, and several thousand molt and stage in eastern Norton Sound. We further identified eastern Norton Sound as the principal molting and staging area for females nesting on the Yukon-Kuskokwim Delta, and Ledyard Bay and Mechigmenskiy Bay as the principal molting and staging areas for females nesting on the North Slope. Males marked at all three breeding grounds molt and stage in Mechigmenskiy Bay, Ledyard Bay, and the Indigirka-Kolyma delta region. Males from the Yukon-Kuskokwim Delta molt and stage mainly at Mechigmenskiy Bay. Equal numbers of males from the North Slope molt and stage at all three areas, and most males from arctic Russia molt and stage at the Indigirka-Kolyma delta region. Postbreeding migration corridors were offshore in the Bering, Chukchi, and Beaufort seas. In winter, eiders were in the Bering Sea south of St. Lawrence Island. Our estimates from surveys in late winter and early spring suggest that at least 333,000 birds winter in single-species flocks in the pack ice in the Bering Sea.
Glackin, Brendan; Wall, Julie A.; McGinnity, Thomas M.; Maguire, Liam P.; McDaid, Liam J.
2010-01-01
Sound localization can be defined as the ability to identify the position of an input sound source and is considered a powerful aspect of mammalian perception. For low frequency sounds, i.e., in the range 270 Hz–1.5 KHz, the mammalian auditory pathway achieves this by extracting the Interaural Time Difference between sound signals being received by the left and right ear. This processing is performed in a region of the brain known as the Medial Superior Olive (MSO). This paper presents a Spiking Neural Network (SNN) based model of the MSO. The network model is trained using the Spike Timing Dependent Plasticity learning rule using experimentally observed Head Related Transfer Function data in an adult domestic cat. The results presented demonstrate how the proposed SNN model is able to perform sound localization with an accuracy of 91.82% when an error tolerance of ±10° is used. For angular resolutions down to 2.5°, it will be demonstrated how software based simulations of the model incur significant computation times. The paper thus also addresses preliminary implementation on a Field Programmable Gate Array based hardware platform to accelerate system performance. PMID:20802855
Directionally Antagonistic Graphene Oxide-Polyurethane Hybrid Aerogel as a Sound Absorber.
Oh, Jung-Hwan; Kim, Jieun; Lee, Hyeongrae; Kang, Yeonjune; Oh, Il-Kwon
2018-06-21
Innovative sound absorbers, the design of which is based on carbon nanotubes and graphene derivatives, could be used to make more efficient sound absorbing materials because of their excellent intrinsic mechanical and chemical properties. However, controlling the directional alignments of low-dimensional carbon nanomaterials, such as restacking, alignment, and dispersion, has been a challenging problem when developing sound absorbing forms. Herein, we present the directionally antagonistic graphene oxide-polyurethane hybrid aerogel we developed as a sound absorber, the physical properties of which differ according to the alignment of the microscopic graphene oxide sheets. This porous graphene sound absorber has a microporous hierarchical cellular structure with adjustable stiffness and improved sound absorption performance, thereby overcoming the restrictions of both geometric and function-orientated functions. Furthermore, by controlling the inner cell size and aligned structure of graphene oxide layers in this study, we achieved remarkable improvement of the sound absorption performance at low frequency. This improvement is attributed to multiple scattering of incident and reflection waves on the aligned porous surfaces, and air-viscous resistance damping inside interconnected structures between the urethane foam and the graphene oxide network. Two anisotropic sound absorbers based on the directionally antagonistic graphene oxide-polyurethane hybrid aerogels were fabricated. They show remarkable differences owing to the opposite alignment of graphene oxide layers inside the polyurethane foam and are expected to be appropriate for the engineering design of sound absorbers in consideration of the wave direction.
Molding acoustic, electromagnetic and water waves with a single cloak
Xu, Jun; Jiang, Xu; Fang, Nicholas; Georget, Elodie; Abdeddaim, Redha; Geffrin, Jean-Michel; Farhat, Mohamed; Sabouroux, Pierre; Enoch, Stefan; Guenneau, Sébastien
2015-01-01
We describe two experiments demonstrating that a cylindrical cloak formerly introduced for linear surface liquid waves works equally well for sound and electromagnetic waves. This structured cloak behaves like an acoustic cloak with an effective anisotropic density and an electromagnetic cloak with an effective anisotropic permittivity, respectively. Measured forward scattering for pressure and magnetic fields are in good agreement and provide first evidence of broadband cloaking. Microwave experiments and 3D electromagnetic wave simulations further confirm reduced forward and backscattering when a rectangular metallic obstacle is surrounded by the structured cloak for cloaking frequencies between 2.6 and 7.0 GHz. This suggests, as supported by 2D finite element simulations, sound waves are cloaked between 3 and 8 KHz and linear surface liquid waves between 5 and 16 Hz. Moreover, microwave experiments show the field is reduced by 10 to 30 dB inside the invisibility region, which suggests the multi-wave cloak could be used as a protection against water, sonic or microwaves. PMID:26057934
Using voice input and audio feedback to enhance the reality of a virtual experience
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miner, N.E.
1994-04-01
Virtual Reality (VR) is a rapidly emerging technology which allows participants to experience a virtual environment through stimulation of the participant`s senses. Intuitive and natural interactions with the virtual world help to create a realistic experience. Typically, a participant is immersed in a virtual environment through the use of a 3-D viewer. Realistic, computer-generated environment models and accurate tracking of a participant`s view are important factors for adding realism to a virtual experience. Stimulating a participant`s sense of sound and providing a natural form of communication for interacting with the virtual world are equally important. This paper discusses the advantagesmore » and importance of incorporating voice recognition and audio feedback capabilities into a virtual world experience. Various approaches and levels of complexity are discussed. Examples of the use of voice and sound are presented through the description of a research application developed in the VR laboratory at Sandia National Laboratories.« less
Ballistic range experiments on superbooms generated by refraction
NASA Technical Reports Server (NTRS)
Sanai, M.; Toong, T.-Y.; Pierce, A. D.
1976-01-01
The enhanced sonic boom or supersonic boom generated as a result of atmospheric refraction in threshold Mach number flights was recreated in a ballistic range by firing projectiles at low supersonic speeds into a stratified medium obtained by slowly injecting carbon dioxide into air. The range was equipped with a fast-response dynamic pressure transducer and schlieren photographic equipment, and the sound speed variation with height was controlled by regulating the flow rate of the CO2. The schlieren observations of the resulting flow field indicate that the generated shocks are reflected near the sonic cutoff altitude where local sound speed equals body speed, provided such an altitude exists. Maximum shock strength occurs very nearly at the point where the incident and reflected shocks join, indicating that the presence of the reflected shock may have an appreciable effect on the magnitude of the focus factor. The largest focus factor detected was 1.7 and leads to an estimate that the constant in the Guiraud-Thery scaling law should have a value of 1.30.
Method for Ultrasonic Imaging and Device for Performing the Method
NASA Technical Reports Server (NTRS)
Madaras, Eric I. (Inventor)
1997-01-01
A method for ultrasonic imaging of interior structures and flaws in a test specimen with a smooth or irregular contact surfaces, in which an ultrasonic transducer is coupled acoustically to the contact surface via a plurality of ultrasonic wave guides with equal delay times. The wave guides are thin and bendable, so they adapt to variations in the distance between the transducer and different parts of the contact surface by bending more or less. All parts of the irregular contact surface accordingly receive sound waves that are in phase, even when the contact surface is irregular, so a coherent sound wave is infused in the test specimen. The wave guides can be arranged in the form of an ultrasonic brush, with a flat head for coupling to a flat transducer, and free bristles that can be pressed against the test specimen. By bevelling the bristle ends at a suitable angle, shear mode waves can be infused into the test specimen from a longitudinal mode transducer.
An Experimental Investigation of the Flow Structure of Supersonic Impinging Jets
NASA Technical Reports Server (NTRS)
Henderson, Brenda; Bridges, James; Wernet, Mark
2002-01-01
An experimental investigation into the jet structure associated with sound production by a supersonic impinging jet is presented. Large plate impinging tones are investigated for a nozzle pressure ratio (NPR) of 4 and nozzle-to-plate spacings between 1 and 5 nozzle exit diameters, where NPR is equal to the ratio of the stagnation pressure to the pressure at the nozzle lip. Results from phase-locked shadowgraph and phase-averaged digital particle image velocimetry (DPIV) studies indicate that, during the oscillation cycle, the Mach disk oscillates axially, a well defined recirculation zone is created in the subsonic impingement region and moves toward the plate, and the compression and expansion regions in the outer supersonic flow move downstream, Sound appears to be generated in the wall jet at approximately 2.6R from the jet axis, where R is the nozzle exit radius. The oscillatory motion in the wall jet is the result of the periodic fluid motion in the near wall region.
Stereo Sound Field Controller Design Using Partial Model Matching on the Frequency Domain
NASA Astrophysics Data System (ADS)
Kumon, Makoto; Miike, Katsuhiro; Eguchi, Kazuki; Mizumoto, Ikuro; Iwai, Zenta
The objective of sound field control is to make the acoustic characteristics of a listening room close to those of the desired system. Conventional methods apply feedforward controllers, such as digital filters, to achieve this objective. However, feedback controllers are also necessary in order to attenuate noise or to compensate the uncertainty of the acoustic characteristics of the listening room. Since acoustic characteristics are well modeled on the frequency domain, it is efficient to design controllers with respect to frequency responses, but it is difficult to design a multi input multi output (MIMO) control system on a wide frequency domain. In the present study, a partial model matching method on the frequency domain was adopted because this method requires only sampled data, rather than complex mathematical models of the plant, in order to design controllers for MIMO systems. The partial model matching method was applied to design two-degree-of-freedom controllers for acoustic equalization and noise reduction. Experiments demonstrated effectiveness of the proposed method.
Molding acoustic, electromagnetic and water waves with a single cloak.
Xu, Jun; Jiang, Xu; Fang, Nicholas; Georget, Elodie; Abdeddaim, Redha; Geffrin, Jean-Michel; Farhat, Mohamed; Sabouroux, Pierre; Enoch, Stefan; Guenneau, Sébastien
2015-06-09
We describe two experiments demonstrating that a cylindrical cloak formerly introduced for linear surface liquid waves works equally well for sound and electromagnetic waves. This structured cloak behaves like an acoustic cloak with an effective anisotropic density and an electromagnetic cloak with an effective anisotropic permittivity, respectively. Measured forward scattering for pressure and magnetic fields are in good agreement and provide first evidence of broadband cloaking. Microwave experiments and 3D electromagnetic wave simulations further confirm reduced forward and backscattering when a rectangular metallic obstacle is surrounded by the structured cloak for cloaking frequencies between 2.6 and 7.0 GHz. This suggests, as supported by 2D finite element simulations, sound waves are cloaked between 3 and 8 KHz and linear surface liquid waves between 5 and 16 Hz. Moreover, microwave experiments show the field is reduced by 10 to 30 dB inside the invisibility region, which suggests the multi-wave cloak could be used as a protection against water, sonic or microwaves.
Sereda, Magdalena; Davies, Jeff; Hall, Deborah A
2017-04-01
This report considers feasibility of conducting a UK trial of combination devices for tinnitus, using data from the study which evaluated different listener programmes available within the pre-market version of Oticon Alta with Tinnitus Sound Generator. Open and closed questions addressed the following feasibility issues: (1) Participant recruitment; (2) Device acceptability; (3) Programme preferences in different self-nominated listening situations; (4) Usability; (5) Compliance; (6) Adverse events. Eight current combination hearing aid users (all males) aged between 62-72 years (mean age 67.25 years, SD = 3.8). All eight participants reported the physical aspects and noise options on the experimental device to be acceptable. Programmes with amplification and masking features were equally preferred over the basic amplification-only programme. Individual preferences for the different programme options varied widely, both across participants and across listening situations. A set of recommendations for future trials were formulated which calls for more "real world" trial design rather than tightly controlling the fitting procedure.
Thermal and viscous effects on sound waves: revised classical theory.
Davis, Anthony M J; Brenner, Howard
2012-11-01
In this paper the recently developed, bi-velocity model of fluid mechanics based on the principles of linear irreversible thermodynamics (LIT) is applied to sound propagation in gases taking account of first-order thermal and viscous dissipation effects. The results are compared and contrasted with the classical Navier-Stokes-Fourier results of Pierce for this same situation cited in his textbook. Comparisons are also made with the recent analyses of Dadzie and Reese, whose molecularly based sound propagation calculations furnish results virtually identical with the purely macroscopic LIT-based bi-velocity results below, as well as being well-supported by experimental data. Illustrative dissipative sound propagation examples involving application of the bi-velocity model to several elementary situations are also provided, showing the disjoint entropy mode and the additional, evanescent viscous mode.
Displaying Composite and Archived Soundings in the Advanced Weather Interactive Processing System
NASA Technical Reports Server (NTRS)
Barrett, Joe H., III; Volkmer, Matthew R.; Blottman, Peter F.; Sharp, David W.
2008-01-01
In a previous task, the Applied Meteorology Unit (AMU) developed spatial and temporal climatologies of lightning occurrence based on eight atmospheric flow regimes. The AMU created climatological, or composite, soundings of wind speed and direction, temperature, and dew point temperature at four rawinsonde observation stations at Jacksonville, Tampa, Miami, and Cape Canaveral Air Force Station, for each of the eight flow regimes. The composite soundings were delivered to the National Weather Service (NWS) Melbourne (MLB) office for display using the National version of the Skew-T Hodograph analysis and Research Program (NSHARP) software program. The NWS MLB requested the AMU make the composite soundings available for display in the Advanced Weather Interactive Processing System (AWIPS), so they could be overlaid on current observed soundings. This will allow the forecasters to compare the current state of the atmosphere with climatology. This presentation describes how the AMU converted the composite soundings from NSHARP Archive format to Network Common Data Form (NetCDF) format, so that the soundings could be displayed in AWl PS. The NetCDF is a set of data formats, programming interfaces, and software libraries used to read and write scientific data files. In AWIPS, each meteorological data type, such as soundings or surface observations, has a unique NetCDF format. Each format is described by a NetCDF template file. Although NetCDF files are in binary format, they can be converted to a text format called network Common data form Description Language (CDL). A software utility called ncgen is used to create a NetCDF file from a CDL file, while the ncdump utility is used to create a CDL file from a NetCDF file. An AWIPS receives soundings in Binary Universal Form for the Representation of Meteorological data (BUFR) format (http://dss.ucar.edu/docs/formats/bufr/), and then decodes them into NetCDF format. Only two sounding files are generated in AWIPS per day. One file contains all of the soundings received worldwide between 0000 UTC and 1200 UTC, and the other includes all soundings between 1200 UTC and 0000 UTC. In order to add the composite soundings into AWIPS, a procedure was created to configure, or localize, AWIPS. This involved modifying and creating several configuration text files. A unique fourcharacter site identifier was created for each of the 32 soundings so each could be viewed separately. The first three characters were based on the site identifier of the observed sounding, while the last character was based on the flow regime. While researching the localization process for soundings, the AMU discovered a method of archiving soundings so old soundings would not get purged automatically by AWl PS. This method could provide an alternative way of localizing AWl PS for composite soundings. In addition, this would allow forecasters to use archived soundings in AWIPS for case studies. A test sounding file in NetCDF format was written in order to verify the correct format for soundings in AWIPS. After the file was viewed successfully in AWIPS, the AMU wrote a software program in the Tool Command Language/Tool Kit (Tcl/Tk) language to convert the 32 composite soundings from NSHARP Archive to CDL format. The ncgen utility was then used to convert the CDL file to a NetCDF file. The NetCDF file could then be read and displayed in AWIPS.
Different Timescales for the Neural Coding of Consonant and Vowel Sounds
Perez, Claudia A.; Engineer, Crystal T.; Jakkamsetti, Vikram; Carraway, Ryan S.; Perry, Matthew S.
2013-01-01
Psychophysical, clinical, and imaging evidence suggests that consonant and vowel sounds have distinct neural representations. This study tests the hypothesis that consonant and vowel sounds are represented on different timescales within the same population of neurons by comparing behavioral discrimination with neural discrimination based on activity recorded in rat inferior colliculus and primary auditory cortex. Performance on 9 vowel discrimination tasks was highly correlated with neural discrimination based on spike count and was not correlated when spike timing was preserved. In contrast, performance on 11 consonant discrimination tasks was highly correlated with neural discrimination when spike timing was preserved and not when spike timing was eliminated. These results suggest that in the early stages of auditory processing, spike count encodes vowel sounds and spike timing encodes consonant sounds. These distinct coding strategies likely contribute to the robust nature of speech sound representations and may help explain some aspects of developmental and acquired speech processing disorders. PMID:22426334
Cross-Modal Correspondences Enhance Performance on a Colour-to-Sound Sensory Substitution Device.
Hamilton-Fletcher, Giles; Wright, Thomas D; Ward, Jamie
Visual sensory substitution devices (SSDs) can represent visual characteristics through distinct patterns of sound, allowing a visually impaired user access to visual information. Previous SSDs have avoided colour and when they do encode colour, have assigned sounds to colour in a largely unprincipled way. This study introduces a new tablet-based SSD termed the ‘Creole’ (so called because it combines tactile scanning with image sonification) and a new algorithm for converting colour to sound that is based on established cross-modal correspondences (intuitive mappings between different sensory dimensions). To test the utility of correspondences, we examined the colour–sound associative memory and object recognition abilities of sighted users who had their device either coded in line with or opposite to sound–colour correspondences. Improved colour memory and reduced colour-errors were made by users who had the correspondence-based mappings. Interestingly, the colour–sound mappings that provided the highest improvements during the associative memory task also saw the greatest gains for recognising realistic objects that also featured these colours, indicating a transfer of abilities from memory to recognition. These users were also marginally better at matching sounds to images varying in luminance, even though luminance was coded identically across the different versions of the device. These findings are discussed with relevance for both colour and correspondences for sensory substitution use.
On the effectiveness of vocal imitations and verbal descriptions of sounds.
Lemaitre, Guillaume; Rocchesso, Davide
2014-02-01
Describing unidentified sounds with words is a frustrating task and vocally imitating them is often a convenient way to address the issue. This article reports on a study that compared the effectiveness of vocal imitations and verbalizations to communicate different referent sounds. The stimuli included mechanical and synthesized sounds and were selected on the basis of participants' confidence in identifying the cause of the sounds, ranging from easy-to-identify to unidentifiable sounds. The study used a selection of vocal imitations and verbalizations deemed adequate descriptions of the referent sounds. These descriptions were used in a nine-alternative forced-choice experiment: Participants listened to a description and picked one sound from a list of nine possible referent sounds. Results showed that recognition based on verbalizations was maximally effective when the referent sounds were identifiable. Recognition accuracy with verbalizations dropped when identifiability of the sounds decreased. Conversely, recognition accuracy with vocal imitations did not depend on the identifiability of the referent sounds and was as high as with the best verbalizations. This shows that vocal imitations are an effective means of representing and communicating sounds and suggests that they could be used in a number of applications.
Scattering of sound by atmospheric turbulence predictions in a refractive shadow zone
NASA Technical Reports Server (NTRS)
Mcbride, Walton E.; Bass, Henry E.; Raspet, Richard; Gilbert, Kenneth E.
1990-01-01
According to ray theory, regions exist in an upward refracting atmosphere where no sound should be present. Experiments show, however, that appreciable sound levels penetrate these so-called shadow zones. Two mechanisms contribute to sound in the shadow zone: diffraction and turbulent scattering of sound. Diffractive effects can be pronounced at lower frequencies but are small at high frequencies. In the short wavelength limit, then, scattering due to turbulence should be the predominant mechanism involved in producing the sound levels measured in shadow zones. No existing analytical method includes turbulence effects in the prediction of sound pressure levels in upward refractive shadow zones. In order to obtain quantitative average sound pressure level predictions, a numerical simulation of the effect of atmospheric turbulence on sound propagation is performed. The simulation is based on scattering from randomly distributed scattering centers ('turbules'). Sound pressure levels are computed for many realizations of a turbulent atmosphere. Predictions from the numerical simulation are compared with existing theories and experimental data.
A Sparsity-Based Approach to 3D Binaural Sound Synthesis Using Time-Frequency Array Processing
NASA Astrophysics Data System (ADS)
Cobos, Maximo; Lopez, JoseJ; Spors, Sascha
2010-12-01
Localization of sounds in physical space plays a very important role in multiple audio-related disciplines, such as music, telecommunications, and audiovisual productions. Binaural recording is the most commonly used method to provide an immersive sound experience by means of headphone reproduction. However, it requires a very specific recording setup using high-fidelity microphones mounted in a dummy head. In this paper, we present a novel processing framework for binaural sound recording and reproduction that avoids the use of dummy heads, which is specially suitable for immersive teleconferencing applications. The method is based on a time-frequency analysis of the spatial properties of the sound picked up by a simple tetrahedral microphone array, assuming source sparseness. The experiments carried out using simulations and a real-time prototype confirm the validity of the proposed approach.
Teaching room acoustics as a product sound quality issue
NASA Astrophysics Data System (ADS)
Kleiner, Mendel; Vastfjall, Daniel
2003-04-01
The department of Applied Acoustics teaches engineering and architect students at Chalmers University of Technology. The teaching of room acoustics to architectural students has been under constant development under several years and is now based on the study of room acoustics as a product sound quality issue. Various listening sessions using binaural sound recording and reproduction is used to focus students' learning on simple, easy to remember concepts. Computer modeling using ray tracing software and auralization is also used extensively as a tool to demonstrate concepts in addition to other software for simple sound generation and manipulation. Sound in general is the focus of an interdisciplinary course for students from Chalmers as well as from a school of art, a school of design, and a school of music which offers particular challenges and which is almost all listening based.
Predicting Hail Size Using Model Vertical Velocities
2008-03-01
updrafts from a simple cloud model using forecasted soundings . The models used MM5 model data coinciding with severe hail events collected from the...updrafts from a simple cloud model using forecasted soundings . The models used MM5 model data coinciding with severe hail events collected from the...determine their accuracy. Plus they are based primary on observed upper air soundings . Obtaining upper air soundings in proximity to convective
A State-of-the-Art Review: Personalization of Tinnitus Sound Therapy.
Searchfield, Grant D; Durai, Mithila; Linford, Tania
2017-01-01
Background: There are several established, and an increasing number of putative, therapies using sound to treat tinnitus. There appear to be few guidelines for sound therapy selection and application. Aim: To review current approaches to personalizing sound therapy for tinnitus. Methods: A "state-of-the-art" review (Grant and Booth, 2009) was undertaken to answer the question: how do current sound-based therapies for tinnitus adjust for tinnitus heterogeneity? Scopus, Google Scholar, Embase and PubMed were searched for the 10-year period 2006-2016. The search strategy used the following key words: "tinnitus" AND "sound" AND "therapy" AND "guidelines" OR "personalized" OR "customized" OR "individual" OR "questionnaire" OR "selection." The results of the review were cataloged and organized into themes. Results: In total 165 articles were reviewed in full, 83 contained sufficient details to contribute to answering the study question. The key themes identified were hearing compensation, pitched-match therapy, maskability, reaction to sound and psychosocial factors. Although many therapies mentioned customization, few could be classified as being personalized. Several psychoacoustic and questionnaire-based methods for assisting treatment selection were identified. Conclusions: Assessment methods are available to assist clinicians to personalize sound-therapy and empower patients to be active in therapy decision-making. Most current therapies are modified using only one characteristic of the individual and/or their tinnitus.
AUDIS wear: a smartwatch based assistive device for ubiquitous awareness of environmental sounds.
Mielke, Matthias; Bruck, Rainer
2016-08-01
A multitude of assistive devices is available for deaf people (i.e. deaf, deafened, and hard of hearing). Besides hearing and communication aids, devices to access environmental sounds are available commercially. But the devices have two major drawbacks: 1. they are targeted at indoor environments (e.g. home or work), and 2. only specific events are supported (e.g. the doorbell or telephone). Recent research shows that important sounds can occur in all contexts and that the interests in sounds are diverse. These drawbacks can be tackled by using modern information and communication technology that enables the development of new and improved assistive devices. The smartwatch, a new computing platform in the form of a wristwatch, offers new potential for assistive technology. Its design promises a perfect integration into various different social contexts and thus blends perfectly into the user's life. Based on a smartwatch and algorithms from pattern recognition, a prototype for awareness of environmental sounds is presented here. It observes the acoustic environment of the user and detects environmental sounds. A vibration is triggered when a sound is detected and the type of sound is shown on the display. The design of the prototype was discussed with deaf people in semi-structured interviews, leading to a set of implications for the design of such a device.
A neurally inspired musical instrument classification system based upon the sound onset.
Newton, Michael J; Smith, Leslie S
2012-06-01
Physiological evidence suggests that sound onset detection in the auditory system may be performed by specialized neurons as early as the cochlear nucleus. Psychoacoustic evidence shows that the sound onset can be important for the recognition of musical sounds. Here the sound onset is used in isolation to form tone descriptors for a musical instrument classification task. The task involves 2085 isolated musical tones from the McGill dataset across five instrument categories. A neurally inspired tone descriptor is created using a model of the auditory system's response to sound onset. A gammatone filterbank and spiking onset detectors, built from dynamic synapses and leaky integrate-and-fire neurons, create parallel spike trains that emphasize the sound onset. These are coded as a descriptor called the onset fingerprint. Classification uses a time-domain neural network, the echo state network. Reference strategies, based upon mel-frequency cepstral coefficients, evaluated either over the whole tone or only during the sound onset, provide context to the method. Classification success rates for the neurally-inspired method are around 75%. The cepstral methods perform between 73% and 76%. Further testing with tones from the Iowa MIS collection shows that the neurally inspired method is considerably more robust when tested with data from an unrelated dataset.
Sound production by singing humpback whales.
Mercado, Eduardo; Schneider, Jennifer N; Pack, Adam A; Herman, Louis M
2010-04-01
Sounds from humpback whale songs were analyzed to evaluate possible mechanisms of sound production. Song sounds fell along a continuum with trains of discrete pulses at one end and continuous tonal signals at the other. This graded vocal repertoire is comparable to that seen in false killer whales [Murray et al. (1998). J. Acoust. Soc. Am. 104, 1679-1688] and human singers, indicating that all three species generate sounds by varying the tension of pneumatically driven, vibrating membranes. Patterns in the spectral content of sounds and in nonlinear sound features show that resonating air chambers may also contribute to humpback whale sound production. Collectively, these findings suggest that categorizing individual units within songs into discrete types may obscure how singers modulate song features and illustrate how production-based characterizations of vocalizations can provide new insights into how humpback whales sing.
The impact of artificial vehicle sounds for pedestrians on driver stress.
Cottrell, Nicholas D; Barton, Benjamin K
2012-01-01
Electrically based vehicles have produced some concern over their lack of sound, but the impact of artificial sounds now being implemented have not been examined in respect to their effects upon the driver. The impact of two different implementations of vehicle sound on driver stress in electric vehicles was examined. A Nissan HEV running in electric vehicle mode was driven by participants in an area of congestion using three sound implementations: (1) no artificial sounds, (2) manually engaged sounds and (3) automatically engaged sounds. Physiological and self-report questionnaire measures were collected to determine stress and acceptance of the automated sound protocol. Driver stress was significantly higher in the manually activated warning condition, compared to both no artificial sounds and automatically engaged sounds. Implications for automation usage and measurement methods are discussed and future research directions suggested. The advent of hybrid- and all-electric vehicles has created a need for artificial warning signals for pedestrian safety that place task demands on drivers. We investigated drivers' stress differences in response to varying conditions of warning signals for pedestrians. Driver stress was lower when noises were automated.
Sound therapy for tinnitus management: practicable options.
Hoare, Derek J; Searchfield, Grant D; El Refaie, Amr; Henry, James A
2014-01-01
The authors reviewed practicable options of sound therapy for tinnitus, the evidence base for each option, and the implications of each option for the patient and for clinical practice. To provide a general guide to selecting sound therapy options in clinical practice. Practicable sound therapy options. Where available, peer-reviewed empirical studies, conference proceedings, and review studies were examined. Material relevant to the purpose was summarized in a narrative. The number of peer-reviewed publications pertaining to each sound therapy option reviewed varied significantly (from none to over 10). Overall there is currently insufficient evidence to support or refute the routine use of individual sound therapy options. It is likely, however, that sound therapy combined with education and counseling is generally helpful to patients. Clinicians need to be guided by the patient's point of care, patient motivation and expectations of sound therapy, and the acceptability of the intervention both in terms of the sound stimuli they are to use and whether they are willing to use sound extensively or intermittently. Clinicians should also clarify to patients the role sound therapy is expected to play in the management plan. American Academy of Audiology.
Psychoacoustical evaluation of natural and urban sounds in soundscapes.
Yang, Ming; Kang, Jian
2013-07-01
Among various sounds in the environment, natural sounds, such as water sounds and birdsongs, have proven to be highly preferred by humans, but the reasons for these preferences have not been thoroughly researched. This paper explores differences between various natural and urban environmental sounds from the viewpoint of objective measures, especially psychoacoustical parameters. The sound samples used in this study include the recordings of single sound source categories of water, wind, birdsongs, and urban sounds including street music, mechanical sounds, and traffic noise. The samples are analyzed with a number of existing psychoacoustical parameter algorithmic models. Based on hierarchical cluster and principal components analyses of the calculated results, a series of differences has been shown among different sound types in terms of key psychoacoustical parameters. While different sound categories cannot be identified using any single acoustical and psychoacoustical parameter, identification can be made with a group of parameters, as analyzed with artificial neural networks and discriminant functions in this paper. For artificial neural networks, correlations between network predictions and targets using the average and standard deviation data of psychoacoustical parameters as inputs are above 0.95 for the three natural sound categories and above 0.90 for the urban sound category. For sound identification/classification, key parameters are fluctuation strength, loudness, and sharpness.
A multichannel amplitude and relative-phase controller for active sound quality control
NASA Astrophysics Data System (ADS)
Mosquera-Sánchez, Jaime A.; Desmet, Wim; de Oliveira, Leopoldo P. R.
2017-05-01
The enhancement of the sound quality of periodic disturbances for a number of listeners within an enclosure often confronts difficulties given by cross-channel interferences, which arise from simultaneously profiling the primary sound at each error sensor. These interferences may deteriorate the original sound among each listener, which is an unacceptable result from the point of view of sound quality control. In this paper we provide experimental evidence on controlling both amplitude and relative-phase functions of stationary complex primary sounds for a number of listeners within a cavity, attaining amplifications of twice the original value, reductions on the order of 70 dB, and relative-phase shifts between ± π rad, still in a free-of-interference control scenario. To accomplish such burdensome control targets, we have designed a multichannel active sound profiling scheme that bases its operation on exchanging time-domain control signals among the control units during uptime. Provided the real parts of the eigenvalues of persistently excited control matrices are positive, the proposed multichannel array is able to counterbalance cross-channel interferences, while attaining demanding control targets. Moreover, regularization of unstable control matrices is not seen to prevent the proposed array to provide free-of-interference amplitude and relative-phase control, but the system performance is degraded, as a function of the amount of regularization needed. The assessment of Loudness and Roughness metrics on the controlled primary sound proves that the proposed distributed control scheme noticeably outperforms current techniques, since active amplitude- and/or relative-phase-based enhancement of the auditory qualities of a primary sound no longer implies in causing interferences among different positions. In this regard, experimental results also confirm the effectiveness of the proposed scheme on stably enhancing the sound qualities of periodic sounds for multiple listeners within a cavity.
Wake Vortex Avoidance System and Method
NASA Technical Reports Server (NTRS)
Shams, Qamar A. (Inventor); Zuckerwar, Allan J. (Inventor); Knight, Howard K. (Inventor)
2017-01-01
A wake vortex avoidance system includes a microphone array configured to detect low frequency sounds. A signal processor determines a geometric mean coherence based on the detected low frequency sounds. A display displays wake vortices based on the determined geometric mean coherence.
Quantitative photoacoustic imaging in the acoustic regime using SPIM
NASA Astrophysics Data System (ADS)
Beigl, Alexander; Elbau, Peter; Sadiq, Kamran; Scherzer, Otmar
2018-05-01
While in standard photoacoustic imaging the propagation of sound waves is modeled by the standard wave equation, our approach is based on a generalized wave equation with variable sound speed and material density, respectively. In this paper we present an approach for photoacoustic imaging, which in addition to the recovery of the absorption density parameter, the imaging parameter of standard photoacoustics, also allows us to reconstruct the spatially varying sound speed and density, respectively, of the medium. We provide analytical reconstruction formulas for all three parameters based in a linearized model based on single plane illumination microscopy (SPIM) techniques.
Development of Prototype of Whistling Sound Counter based on Piezoelectric Bone Conduction
NASA Astrophysics Data System (ADS)
Mori, Mikio; Ogihara, Mitsuhiro; Kyuu, Ten; Taniguchi, Shuji; Kato, Shozo; Araki, Chikahiro
Recently, some professional whistlers have set up music schools that teach musical whistling. Similar to singing, in musical whistling, the whistling sound should not be break, even when the whistling goes on for more than 3 min. For this, it is advisable to practice whistling the “Pii” sound, which involves whistling the “Pii” sound continuously 100 times with the same pitch. However, when practicing alone, a whistler finds it difficult to count his/her own whistling sounds. In this paper, we propose a whistling sound counter based on piezoelectric bone conduction. This system consists of five parts. The gain of the amplifier section of this counter is variable, and the center frequency (f0) of the BPF part is also variable. In this study, we developed a prototype of the system and tested it. For this, we simultaneously counted the whistling sounds of nine people using the proposed system. The proposed system showed a good performance in a noisy environment. We also propose an examination system for awarding grades in musical whistling, which enforces the license examination in musical whistling on the personal computer. The proposed system can be used to administer the 5th grade exam for musical whistling.
Health care reform and people with disabilities.
Batavia, A I
1993-01-01
As a group, people with disabilities or chronic conditions experience higher-than-average health care costs and have difficulty gaining access to affordable private health insurance coverage. While the Americans with Disabilities Act will enhance access by prohibiting differential treatment without sound actuarial justification, it will not guarantee equal access for people in impairment groups with high utilization rates. Health care reform is needed to subsidize the coverage of such individuals. Such subsidization can be achieved under either a casualty insurance model, in which premiums based on expected costs are subsidized directly, or a social insurance model, in which low-cost enrollees cross-subsidize high-cost enrollees. Cost containment provisions that focus on the provider, such as global budgeting and managed competition, will adversely affect disabled people if providers do not have adequate incentives to meet these people's needs. Provisions focusing on the consumer, such as cost sharing, case management, and benefit reductions, will adversely affect disabled people if they unduly limit needed services or impose a disproportionate financial burden on disabled people.
Deep-Focusing Time-Distance Helioseismology
NASA Technical Reports Server (NTRS)
Duvall, T. L., Jr.; Jensen, J. M.; Kosovichev, A. G.; Birch, A. C.; Fisher, Richard R. (Technical Monitor)
2001-01-01
Much progress has been made by measuring the travel times of solar acoustic waves from a central surface location to points at equal arc distance away. Depth information is obtained from the range of arc distances examined, with the larger distances revealing the deeper layers. This method we will call surface-focusing, as the common point, or focus, is at the surface. To obtain a clearer picture of the subsurface region, it would, no doubt, be better to focus on points below the surface. Our first attempt to do this used the ray theory to pick surface location pairs that would focus on a particular subsurface point. This is not the ideal procedure, as Born approximation kernels suggest that this focus should have zero sensitivity to sound speed inhomogeneities. However, the sensitivity is concentrated below the surface in a much better way than the old surface-focusing method, and so we expect the deep-focusing method to be more sensitive. A large sunspot group was studied by both methods. Inversions based on both methods will be compared.
Real-time dual-band haptic music player for mobile devices.
Hwang, Inwook; Lee, Hyeseon; Choi, Seungmoon
2013-01-01
We introduce a novel dual-band haptic music player for real-time simultaneous vibrotactile playback with music in mobile devices. Our haptic music player features a new miniature dual-mode actuator that can produce vibrations consisting of two principal frequencies and a real-time vibration generation algorithm that can extract vibration commands from a music file for dual-band playback (bass and treble). The algorithm uses a "haptic equalizer" and provides plausible sound-to-touch modality conversion based on human perceptual data. In addition, we present a user study carried out to evaluate the subjective performance (precision, harmony, fun, and preference) of the haptic music player, in comparison with the current practice of bass-band-only vibrotactile playback via a single-frequency voice-coil actuator. The evaluation results indicated that the new dual-band playback outperforms the bass-only rendering, also providing several insights for further improvements. The developed system and experimental findings have implications for improving the multimedia experience with mobile devices.
A critical review of noise production models for turbulent, gas-fueled burners
NASA Technical Reports Server (NTRS)
Mahan, J. R.
1984-01-01
The combustion noise literature for the period between 1952 and early 1984 is critically reviewed. Primary emphasis is placed on past theoretical and semi-empirical attempts to predict or explain observed direct combustion noise characteristics of turbulent, gas-fueled burners; works involving liquid-fueled burners are reviewed only when ideas equally applicable to gas-fueled burners are pesented. The historical development of the most important contemporary direct combustion noise theories is traced, and the theories themselves are compared and criticized. While most theories explain combustion noise production by turbulent flames in terms of randomly distributed acoustic monopoles produced by turbulent mixing of products and reactants, none is able to predict the sound pressure in the acoustic farfield of a practical burner because of the lack of a proven model which relates the combustion noise source strenght at a given frequency to the design and operating parameters of the burner. Recommendations are given for establishing a benchmark-quality data base needed to support the development of such a model.
Investigation of the effect of pressure increasing in condensing heat-exchanger
NASA Astrophysics Data System (ADS)
Murmanskii, I. B.; Aronson, K. E.; Brodov, Yu M.; Galperin, L. G.; Ryabchikov, A. Yu.; Brezgin, D. V.
2017-11-01
The effect of pressure increase was observed in steam condensation in the intermediate coolers of multistage steam ejector. Steam pressure increase for ejector cooler amounts up to 1.5 kPa in the first ejector stage, 5 kPa in the second and 7 kPa in the third one. Pressure ratios are equal to 2.0, 1.3 and 1.1 respectively. As a rule steam velocities at the cooler inlets do not exceed 40…100 m/s and are subsonic in all regimes. The report presents a computational model that describes the effect of pressure increase in the cooler. The steam entering the heat exchanger tears the drops from the condensate film flowing down vertical tubes. At the inlet of heat exchanger the steam flow capturing condensate droplets forms a steam-water mixture in which the sound velocity is significantly reduced. If the flow rate of steam-water mixture in heat exchanger is greater than the sound velocity, there occurs a pressure shock in the wet steam. On the basis of the equations of mass, momentum and energy conservation the authors derived the expressions for calculation of steam flow dryness degree before and after the shock. The model assumes that droplet velocity is close to the velocity of the steam phase (slipping is absent); drops do not come into thermal interaction with the steam phase; liquid phase specific volume compared to the volume of steam is neglected; pressure shock is calculated taking into account the gas-dynamic flow resistance of the tube bundle. It is also assumed that the temperature of steam after the shock is equal to the saturation temperature. The calculations have shown that the rise of steam pressure and temperature in the shock results in dryness degree increase. For calculated flow parameters the velocity value before the shock is greater than the sound velocity. Thus, on the basis of generally accepted physics knowledge the computational model has been formulated for the effect of steam pressure rise in the condensing heat exchanger.
Sahu, Atanu; Bhattacharya, Partha; Niyogi, Arup Guha; Rose, Michael
2017-06-01
Double-wall panels are known for their superior sound insulation properties over single wall panels as a sound barrier. The sound transmission phenomenon through a double-wall structure is a complex process involving vibroacoustic interaction between structural panels, the air-cushion in between, and the secondary acoustic domain. It is in this context a versatile and a fully coupled technique based on the finite-element-boundary element model is developed that enables estimation of sound transfer through a double-wall panel into an adjacent enclosure while satisfying the displacement compatibility across the interface. The contribution of individual components in the transmitted energy is identified through numerical simulations.
ERIC Educational Resources Information Center
Burns, Gary
Based on the argument that (contrary to critical opinion) the musicians in the various bands associated with Bosstown Sound were indeed talented, cohesive individuals and that the bands' lack of renown was partially a result of ill-treatment by record companies and the press, this paper traces the development of the Bosstown Sound from its…
Moore, Brian C J; Füllgrabe, Christian; Stone, Michael A
2011-01-01
To determine preferred parameters of multichannel compression using individually fitted simulated hearing aids and a method of paired comparisons. Fourteen participants with mild to moderate hearing loss listened via a simulated five-channel compression hearing aid fitted using the CAMEQ2-HF method to pairs of speech sounds (a male talker and a female talker) and musical sounds (a percussion instrument, orchestral classical music, and a jazz trio) presented sequentially and indicated which sound of the pair was preferred and by how much. The sounds in each pair were derived from the same token and differed along a single dimension in the type of processing applied. For the speech sounds, participants judged either pleasantness or clarity; in the latter case, the speech was presented in noise at a 2-dB signal-to-noise ratio. For musical sounds, they judged pleasantness. The parameters explored were time delay of the audio signal relative to the gain control signal (the alignment delay), compression speed (attack and release times), bandwidth (5, 7.5, or 10 kHz), and gain at high frequencies relative to that prescribed by CAMEQ2-HF. Pleasantness increased with increasing alignment delay only for the percussive musical sound. Clarity was not affected by alignment delay. There was a trend for pleasantness to decrease slightly with increasing bandwidth, but this was significant only for female speech with fast compression. Judged clarity was significantly higher for the 7.5- and 10-kHz bandwidths than for the 5-kHz bandwidth for both slow and fast compression and for both talker genders. Compression speed had little effect on pleasantness for 50- or 65-dB SPL input levels, but slow compression was generally judged as slightly more pleasant than fast compression for an 80-dB SPL input level. Clarity was higher for slow than for fast compression for input levels of 80 and 65 dB SPL but not for a level of 50 dB SPL. Preferences for pleasantness were approximately equal with CAMEQ2-HF gains and with gains slightly reduced at high frequencies and were lower when gains were slightly increased at high frequencies. Speech clarity was not affected by changing the gain at high frequencies. Effects of alignment delay were small except for the percussive sound. A wider bandwidth was slightly preferred for speech clarity. Speech clarity was slightly greater with slow compression, especially at high levels. Preferred high-frequency gains were close to or a little below those prescribed by CAMEQ2-HF.
A visual stethoscope to detect the position of the tracheal tube.
Kato, Hiromi; Suzuki, Akira; Nakajima, Yoshiki; Makino, Hiroshi; Sanjo, Yoshimitsu; Nakai, Takayoshi; Shiraishi, Yoshito; Katoh, Takasumi; Sato, Shigehito
2009-12-01
Advancing a tracheal tube into the bronchus produces unilateral breath sounds. We created a Visual Stethoscope that allows real-time fast Fourier transformation of the sound signal and 3-dimensional (frequency-amplitude-time) color rendering of the results on a personal computer with simultaneous processing of 2 individual sound signals. The aim of this study was to evaluate whether the Visual Stethoscope can detect bronchial intubation in comparison with auscultation. After induction of general anesthesia, the trachea was intubated with a tracheal tube. The distance from the incisors to the carina was measured using a fiberoptic bronchoscope. While the anesthesiologist advanced the tracheal tube from the trachea to the bronchus, another anesthesiologist auscultated breath sounds to detect changes of the breath sounds and/or disappearance of bilateral breath sounds for every 1 cm that the tracheal tube was advanced. Two precordial stethoscopes placed at the left and right sides of the chest were used to record breath sounds simultaneously. Subsequently, at a later date, we randomly entered the recorded breath sounds into the Visual Stethoscope. The same anesthesiologist observed the visualized breath sounds on the personal computer screen processed by the Visual Stethoscope to examine changes of breath sounds and/or disappearance of bilateral breath sound. We compared the decision made based on auscultation with that made based on the results of the visualized breath sounds using the Visual Stethoscope. Thirty patients were enrolled in the study. When irregular breath sounds were auscultated, the tip of the tracheal tube was located at 0.6 +/- 1.2 cm on the bronchial side of the carina. Using the Visual Stethoscope, when there were any changes of the shape of the visualized breath sound, the tube was located at 0.4 +/- 0.8 cm on the tracheal side of the carina (P < 0.01). When unilateral breath sounds were auscultated, the tube was located at 2.6 +/- 1.2 cm on the bronchial side of the carina. The tube was also located at 2.3 +/- 1.0 cm on the bronchial side of the carina when a unilateral shape of visualized breath sounds was obtained using the Visual Stethoscope (not significant). During advancement of the tracheal tube, alterations of the shape of the visualized breath sounds using the Visual Stethoscope appeared before the changes of the breath sounds were detected by auscultation. Bilateral breath sounds disappeared when the tip of the tracheal tube was advanced beyond the carina in both groups.
A Low Cost GPS System for Real-Time Tracking of Sounding Rockets
NASA Technical Reports Server (NTRS)
Markgraf, M.; Montenbruck, O.; Hassenpflug, F.; Turner, P.; Bull, B.; Bauer, Frank (Technical Monitor)
2001-01-01
In an effort to minimize the need for costly, complex, tracking radars, the German Space Operations Center has set up a research project for GPS based tracking of sounding rockets. As part of this project, a GPS receiver based on commercial technology for terrestrial applications has been modified to allow its use under the highly dynamical conditions of a sounding rocket flight. In addition, new antenna concepts are studied as an alternative to proven but costly wrap-around antennas.
Optimization of orthotropic distributed-mode loudspeaker using attached masses and multi-exciters.
Lu, Guochao; Shen, Yong; Liu, Ziyun
2012-02-01
Based on the orthotropic model of the plate, the method to optimize the sound response of the distributed-mode loudspeaker (DML) using the attached masses and the multi-exciters has been investigated. The attached masses method will rebuild the modes distribution of the plate, based on which multi-exciter method will smooth the sound response. The results indicate that the method can be used to optimize the sound response of the DML. © 2012 Acoustical Society of America
Airborne sound transmission loss characteristics of wood-frame construction
NASA Astrophysics Data System (ADS)
Rudder, F. F., Jr.
1985-03-01
This report summarizes the available data on the airborne sound transmission loss properties of wood-frame construction and evaluates the methods for predicting the airborne sound transmission loss. The first part of the report comprises a summary of sound transmission loss data for wood-frame interior walls and floor-ceiling construction. Data bases describing the sound transmission loss characteristics of other building components, such as windows and doors, are discussed. The second part of the report presents the prediction of the sound transmission loss of wood-frame construction. Appropriate calculation methods are described both for single-panel and for double-panel construction with sound absorption material in the cavity. With available methods, single-panel construction and double-panel construction with the panels connected by studs may be adequately characterized. Technical appendices are included that summarize laboratory measurements, compare measurement with theory, describe details of the prediction methods, and present sound transmission loss data for common building materials.
Light-induced vibration in the hearing organ
Ren, Tianying; He, Wenxuan; Li, Yizeng; Grosh, Karl; Fridberger, Anders
2014-01-01
The exceptional sensitivity of mammalian hearing organs is attributed to an active process, where force produced by sensory cells boost sound-induced vibrations, making soft sounds audible. This process is thought to be local, with each section of the hearing organ capable of amplifying sound-evoked movement, and nearly instantaneous, since amplification can work for sounds at frequencies up to 100 kHz in some species. To test these fundamental precepts, we developed a method for focally stimulating the living hearing organ with light. Light pulses caused intense and highly damped mechanical responses followed by traveling waves that developed with considerable delay. The delayed response was identical to movements evoked by click-like sounds. This shows that the active process is neither local nor instantaneous, but requires mechanical waves traveling from the cochlear base toward its apex. A physiologically-based mathematical model shows that such waves engage the active process, enhancing hearing sensitivity. PMID:25087606
Real time sound analysis for medical remote monitoring.
Istrate, Dan; Binet, Morgan; Cheng, Sreng
2008-01-01
The increase of aging population in Europe involves more people living alone at home with an increased risk of home accidents or falls. In order to prevent or detect a distress situation in the case of an elderly people living alone, a remote monitoring system based on the sound environment analysis can be used. We have already proposed a system which monitors the sound environment, identifies everyday life sounds and distress expressions in order to participate to an alarm decision. This first system uses a classical sound card on a PC or embedded PC allowing only one channel monitor. In this paper, we propose a new architecture of the remote monitoring system, which relies on a real time multichannel implementation based on an USB acquisition card. This structure allows monitoring eight channels in order to cover all the rooms of an apartment. More than that, the SNR estimation leads currently to the adaptation of the recognition models to environment.
Expertise with artificial non-speech sounds recruits speech-sensitive cortical regions
Leech, Robert; Holt, Lori L.; Devlin, Joseph T.; Dick, Frederic
2009-01-01
Regions of the human temporal lobe show greater activation for speech than for other sounds. These differences may reflect intrinsically specialized domain-specific adaptations for processing speech, or they may be driven by the significant expertise we have in listening to the speech signal. To test the expertise hypothesis, we used a video-game-based paradigm that tacitly trained listeners to categorize acoustically complex, artificial non-linguistic sounds. Before and after training, we used functional MRI to measure how expertise with these sounds modulated temporal lobe activation. Participants’ ability to explicitly categorize the non-speech sounds predicted the change in pre- to post-training activation in speech-sensitive regions of the left posterior superior temporal sulcus, suggesting that emergent auditory expertise may help drive this functional regionalization. Thus, seemingly domain-specific patterns of neural activation in higher cortical regions may be driven in part by experience-based restructuring of high-dimensional perceptual space. PMID:19386919
Investigation of the sound generation mechanisms for in-duct orifice plates.
Tao, Fuyang; Joseph, Phillip; Zhang, Xin; Stalnov, Oksana; Siercke, Matthias; Scheel, Henning
2017-08-01
Sound generation due to an orifice plate in a hard-walled flow duct which is commonly used in air distribution systems (ADS) and flow meters is investigated. The aim is to provide an understanding of this noise generation mechanism based on measurements of the source pressure distribution over the orifice plate. A simple model based on Curle's acoustic analogy is described that relates the broadband in-duct sound field to the surface pressure cross spectrum on both sides of the orifice plate. This work describes careful measurements of the surface pressure cross spectrum over the orifice plate from which the surface pressure distribution and correlation length is deduced. This information is then used to predict the radiated in-duct sound field. Agreement within 3 dB between the predicted and directly measured sound fields is obtained, providing direct confirmation that the surface pressure fluctuations acting over the orifice plates are the main noise sources. Based on the developed model, the contributions to the sound field from different radial locations of the orifice plate are calculated. The surface pressure is shown to follow a U 3.9 velocity scaling law and the area over which the surface sources are correlated follows a U 1.8 velocity scaling law.
Doutres, Olivier; Atalla, Noureddine; Osman, Haisam
2015-06-01
Porous materials are widely used for improving sound absorption and sound transmission loss of vibrating structures. However, their efficiency is limited to medium and high frequencies of sound. A solution for improving their low frequency behavior while keeping an acceptable thickness is to embed resonant structures such as Helmholtz resonators (HRs). This work investigates the absorption and transmission acoustic performances of a cellular porous material with a two-dimensional periodic arrangement of HR inclusions. A low frequency model of a resonant periodic unit cell based on the parallel transfer matrix method is presented. The model is validated by comparison with impedance tube measurements and simulations based on both the finite element method and a homogenization based model. At the HR resonance frequency (i) the transmission loss is greatly improved and (ii) the sound absorption of the foam can be either decreased or improved depending on the HR tuning frequency and on the thickness and properties of the host foam. Finally, the diffuse field sound absorption and diffuse field sound transmission loss performance of a 2.6 m(2) resonant cellular material are measured. It is shown that the improvements observed at the Helmholtz resonant frequency on a single cell are confirmed at a larger scale.
Teaching letter sounds to kindergarten English language learners using incremental rehearsal.
Peterson, Meredith; Brandes, Dana; Kunkel, Amy; Wilson, Jennifer; Rahn, Naomi L; Egan, Andrea; McComas, Jennifer
2014-02-01
Proficiency in letter-sound correspondence is important for decoding connected text. This study examined the effects of an evidence-based intervention, incremental rehearsal (IR), on the letter-sound expression of three kindergarten English language learners (ELLs) performing below the district benchmark for letter-sound fluency. Participants were native speakers of Hmong, Spanish, and Polish. A multiple-baseline design across sets of unknown letter sounds was used to evaluate the effects of IR on letter-sound expression. Visual analysis of the data showed an increase in level and trend when IR was introduced in each phase. Percentage of all non-overlapping data (PAND) ranged from 95% to 100%. All participants exceeded expected growth and reached the spring district benchmark for letter-sound fluency. Results suggest that IR is a promising intervention for increasing letter-sound expression for ELLs who evidence delays in acquiring letter sounds. Copyright © 2013 Society for the Study of School Psychology. Published by Elsevier Ltd. All rights reserved.
Relationship Between Speed of Sound in and Density of Normal and Diseased Rat Livers
NASA Astrophysics Data System (ADS)
Hachiya, Hiroyuki; Ohtsuki, Shigeo; Tanaka, Motonao
1994-05-01
Speed of sound is an important acoustic parameter for quantitative characterization of living tissues. In this paper, the relationship between speed of sound in and density of rat liver tissues are investigated. The speed of sound was measured by the nondeformable technique based on frequency-time analysis of a 3.5 MHz pulse response. The speed of sound in normal livers varied minimally between individuals and was not related to body weight or age. In liver tissues which were administered CCl4, the speed of sound was lower than the speed of sound in normal tissues. The relationship between speed of sound and density in normal, fatty and cirrhotic livers can be fitted well on the line which is estimated using the immiscible liquid model assuming a mixture of normal liver and fat tissues. For 3.5 MHz ultrasound, it is considered that the speed of sound in fresh liver with fatty degeneration is responsible for the fat content and is not strongly dependent on the degree of fibrosis.
Sound level exposure of high-risk infants in different environmental conditions.
Byers, Jacqueline F; Waugh, W Randolph; Lowman, Linda B
2006-01-01
To provide descriptive information about the sound levels to which high-risk infants are exposed in various actual environmental conditions in the NICU, including the impact of physical renovation on sound levels, and to assess the contributions of various types of equipment, alarms, and activities to sound levels in simulated conditions in the NICU. Descriptive and comparative design. Convenience sample of 134 infants at a southeastern quarternary children's hospital. A-weighted decibel (dBA) sound levels under various actual and simulated environmental conditions. The renovated NICU was, on average, 4-6 dBA quieter across all environmental conditions than a comparable nonrenovated room, representing a significant sound level reduction. Sound levels remained above consensus recommendations despite physical redesign and staff training. Respiratory therapy equipment, alarms, staff talking, and infant fussiness contributed to higher sound levels. Evidence-based sound-reducing strategies are proposed. Findings were used to plan environment management as part of a developmental, family-centered care, performance improvement program and in new NICU planning.
Separation and reconstruction of high pressure water-jet reflective sound signal based on ICA
NASA Astrophysics Data System (ADS)
Yang, Hongtao; Sun, Yuling; Li, Meng; Zhang, Dongsu; Wu, Tianfeng
2011-12-01
The impact of high pressure water-jet on the different materials target will produce different reflective mixed sound. In order to reconstruct the reflective sound signals distribution on the linear detecting line accurately and to separate the environment noise effectively, the mixed sound signals acquired by linear mike array were processed by ICA. The basic principle of ICA and algorithm of FASTICA were described in detail. The emulation experiment was designed. The environment noise signal was simulated by using band-limited white noise and the reflective sound signal was simulated by using pulse signal. The reflective sound signal attenuation produced by the different distance transmission was simulated by weighting the sound signal with different contingencies. The mixed sound signals acquired by linear mike array were synthesized by using the above simulated signals and were whitened and separated by ICA. The final results verified that the environment noise separation and the reconstruction of the detecting-line sound distribution can be realized effectively.
Development and Current Status of the “Cambridge” Loudness Models
2014-01-01
This article reviews the evolution of a series of models of loudness developed in Cambridge, UK. The first model, applicable to stationary sounds, was based on modifications of the model developed by Zwicker, including the introduction of a filter to allow for the effects of transfer of sound through the outer and middle ear prior to the calculation of an excitation pattern, and changes in the way that the excitation pattern was calculated. Later, modifications were introduced to the assumed middle-ear transfer function and to the way that specific loudness was calculated from excitation level. These modifications led to a finite calculated loudness at absolute threshold, which made it possible to predict accurately the absolute thresholds of broadband and narrowband sounds, based on the assumption that the absolute threshold corresponds to a fixed small loudness. The model was also modified to give predictions of partial loudness—the loudness of one sound in the presence of another. This allowed predictions of masked thresholds based on the assumption that the masked threshold corresponds to a fixed small partial loudness. Versions of the model for time-varying sounds were developed, which allowed prediction of the masked threshold of any sound in a background of any other sound. More recent extensions incorporate binaural processing to account for the summation of loudness across ears. In parallel, versions of the model for predicting loudness for hearing-impaired ears have been developed and have been applied to the development of methods for fitting multichannel compression hearing aids. PMID:25315375
Visual Presentation Effects on Identification of Multiple Environmental Sounds
Masakura, Yuko; Ichikawa, Makoto; Shimono, Koichi; Nakatsuka, Reio
2016-01-01
This study examined how the contents and timing of a visual stimulus affect the identification of mixed sounds recorded in a daily life environment. For experiments, we presented four environment sounds as auditory stimuli for 5 s along with a picture or a written word as a visual stimulus that might or might not denote the source of one of the four sounds. Three conditions of temporal relations between the visual stimuli and sounds were used. The visual stimulus was presented either: (a) for 5 s simultaneously with the sound; (b) for 5 s, 1 s before the sound (SOA between the audio and visual stimuli was 6 s); or (c) for 33 ms, 1 s before the sound (SOA was 1033 ms). Participants reported all identifiable sounds for those audio–visual stimuli. To characterize the effects of visual stimuli on sound identification, the following were used: the identification rates of sounds for which the visual stimulus denoted its sound source, the rates of other sounds for which the visual stimulus did not denote the sound source, and the frequency of false hearing of a sound that was not presented for each sound set. Results of the four experiments demonstrated that a picture or a written word promoted identification of the sound when it was related to the sound, particularly when the visual stimulus was presented for 5 s simultaneously with the sounds. However, a visual stimulus preceding the sounds had a benefit only for the picture, not for the written word. Furthermore, presentation with a picture denoting a sound simultaneously with the sound reduced the frequency of false hearing. These results suggest three ways that presenting a visual stimulus affects identification of the auditory stimulus. First, activation of the visual representation extracted directly from the picture promotes identification of the denoted sound and suppresses the processing of sounds for which the visual stimulus did not denote the sound source. Second, effects based on processing of the conceptual information promote identification of the denoted sound and suppress the processing of sounds for which the visual stimulus did not denote the sound source. Third, processing of the concurrent visual representation suppresses false hearing. PMID:26973478
The silent base flow and the sound sources in a laminar jet.
Sinayoko, Samuel; Agarwal, Anurag
2012-03-01
An algorithm to compute the silent base flow sources of sound in a jet is introduced. The algorithm is based on spatiotemporal filtering of the flow field and is applicable to multifrequency sources. It is applied to an axisymmetric laminar jet and the resulting sources are validated successfully. The sources are compared to those obtained from two classical acoustic analogies, based on quiescent and time-averaged base flows. The comparison demonstrates how the silent base flow sources shed light on the sound generation process. It is shown that the dominant source mechanism in the axisymmetric laminar jet is "shear-noise," which is a linear mechanism. The algorithm presented here could be applied to fully turbulent flows to understand the aerodynamic noise-generation mechanism. © 2012 Acoustical Society of America
Statistical Analysis for Subjective and Objective Evaluations of Dental Drill Sounds.
Yamada, Tomomi; Kuwano, Sonoko; Ebisu, Shigeyuki; Hayashi, Mikako
2016-01-01
The sound produced by a dental air turbine handpiece (dental drill) can markedly influence the sound environment in a dental clinic. Indeed, many patients report that the sound of a dental drill elicits an unpleasant feeling. Although several manufacturers have attempted to reduce the sound pressure levels produced by dental drills during idling based on ISO 14457, the sound emitted by such drills under active drilling conditions may negatively influence the dental clinic sound environment. The physical metrics related to the unpleasant impressions associated with dental drill sounds have not been determined. In the present study, psychological measurements of dental drill sounds were conducted with the aim of facilitating improvement of the sound environment at dental clinics. Specifically, we examined the impressions elicited by the sounds of 12 types of dental drills in idling and drilling conditions using a semantic differential. The analysis revealed that the impressions of dental drill sounds varied considerably between idling and drilling conditions and among the examined drills. This finding suggests that measuring the sound of a dental drill in idling conditions alone may be insufficient for evaluating the effects of the sound. We related the results of the psychological evaluations to those of measurements of the physical metrics of equivalent continuous A-weighted sound pressure levels (LAeq) and sharpness. Factor analysis indicated that impressions of the dental drill sounds consisted of two factors: "metallic and unpleasant" and "powerful". LAeq had a strong relationship with "powerful impression", calculated sharpness was positively related to "metallic impression", and "unpleasant impression" was predicted by the combination of both LAeq and calculated sharpness. The present analyses indicate that, in addition to a reduction in sound pressure level, refining the frequency components of dental drill sounds is important for creating a comfortable sound environment in dental clinics.
Kotagal, Meera; Richards, Morgan K; Chapman, Teresa; Finch, Lisa; McCann, Bessie; Ormazabal, Amaya; Rush, Robert J; Goldin, Adam B
2015-05-01
Safety concerns about the use of radiation-based imaging such as computed tomography (CT) in children have resulted in national recommendations to use ultrasound (US) for the diagnosis of appendicitis when possible. We evaluated the trends in CT and US use in a statewide sample and the accuracy of these modalities. Patients less than or equal to 18 years undergoing appendectomy in Washington State from 2008 to 2013 were evaluated for preoperative US/CT use, as well as imaging/pathology concordance using data from the Surgical Care and Outcomes Assessment Program. Among 3,353 children, 98.3% underwent preoperative imaging. There was a significant increase in the use of US first over the study period (P < .001). The use of CT at any time during the evaluation decreased. Despite this, in 2013, over 40% of the children still underwent CT imaging. Concordance between US imaging and pathology varied between 40% and 75% at hospitals performing greater than or equal to 10 appendectomies in 2013. Over one third (34.9%) of CT scans performed in the evaluation of children with appendicitis were performed after an indeterminate US. Although the use of US as the first imaging modality to diagnose pediatric appendicitis has increased over the past 5 years, over 40% of children still undergo a CT scan during their preoperative evaluation. Causality for this persistence of CT use is unclear, but could include variability in US accuracy, lack of training, and lack of awareness of the risks of radiation-based imaging. Developing a campaign to focus on continued reduction in CT and increased use of high-quality US should be pursued. Copyright © 2015 Elsevier Inc. All rights reserved.
Climatology of equatorial stratosphere over Lagos, Nigeria
NASA Astrophysics Data System (ADS)
Oyekola, Oyedemi Samuel
We have used 12 complete calendar years (January 1993-December 2004) of monthly averages of measurements made by the Dobson spectrophotometer instrument over an urban site, Lagos (6.6oN, 3.3oE), southwest Nigeria, to study equatorial stratospheric column ozone variations and trends. Our results indicate that the time-averaged total column ozone has a seasonal cy-cle, which maximizes in June and July with a value of 259 Dobson units (DU) and minimizes in February with a magnitude of 250 DU. Statistical analysis of the climatological mean monthly total Dobson O3 record for 1993-2004 show that the local trend is approximately +0.041±0.0011 DU/year (+0.49±0.013% per decade). Spectral analysis was applied to the monthly averages series. The significant periodicity at 95% confidence level demonstrate prominent spectra peaks near 1.9 and 3.6 years, representative of quasi-biennial oscillation (QBO) and quasi-triennial oscillation (QTO), respectively. Signal due to semiannual variation is also identified at Lagos sounding site. Comparison with the ozone observations from Total Ozone Mapping Spectrom-eter (TOMS) on board the Earth-Probe (EP) satellite for the period from 1997 to 2002 reveal that EP/TOMS instrument consistently larger than the ground-based measurement from Dob-son station. Percentage mean relative disparity ranges from -11% to 15%. The root mean square error (RMSE) between satellite and ground-based observations over Lagos ranges be-tween ˜35-83 DU with largest and lowest variability occurring during the ascending phase of solar activity (1999, 10.7 cm radio flux, F10.7 equals 154 flux units) and during the peak phase of solar activity (2001, F10.7 equals 181), respectively.
Threshold for onset of injury in Chinook salmon from exposure to impulsive pile driving sounds.
Halvorsen, Michele B; Casper, Brandon M; Woodley, Christa M; Carlson, Thomas J; Popper, Arthur N
2012-01-01
The risk of effects to fishes and other aquatic life from impulsive sound produced by activities such as pile driving and seismic exploration is increasing throughout the world, particularly with the increased exploitation of oceans for energy production. At the same time, there are few data that provide insight into the effects of these sounds on fishes. The goal of this study was to provide quantitative data to define the levels of impulsive sound that could result in the onset of barotrauma to fish. A High Intensity Controlled Impedance Fluid filled wave Tube was developed that enabled laboratory simulation of high-energy impulsive sound that were characteristic of aquatic far-field, plane-wave acoustic conditions. The sounds used were based upon the impulsive sounds generated by an impact hammer striking a steel shell pile. Neutrally buoyant juvenile Chinook salmon (Oncorhynchus tshawytscha) were exposed to impulsive sounds and subsequently evaluated for barotrauma injuries. Observed injuries ranged from mild hematomas at the lowest sound exposure levels to organ hemorrhage at the highest sound exposure levels. Frequency of observed injuries were used to compute a biological response weighted index (RWI) to evaluate the physiological impact of injuries at the different exposure levels. As single strike and cumulative sound exposure levels (SEL(ss), SEL(cum) respectively) increased, RWI values increased. Based on the results, tissue damage associated with adverse physiological costs occurred when the RWI was greater than 2. In terms of sound exposure levels a RWI of 2 was achieved for 1920 strikes by 177 dB re 1 µPa(2)⋅s SEL(ss) yielding a SEL(cum) of 210 dB re 1 µPa(2)⋅s, and for 960 strikes by 180 dB re 1 µPa(2)⋅s SEL(ss) yielding a SEL(cum) of 210 dB re 1 µPa(2)⋅s. These metrics define thresholds for onset of injury in juvenile Chinook salmon.
Threshold for Onset of Injury in Chinook Salmon from Exposure to Impulsive Pile Driving Sounds
Halvorsen, Michele B.; Casper, Brandon M.; Woodley, Christa M.; Carlson, Thomas J.; Popper, Arthur N.
2012-01-01
The risk of effects to fishes and other aquatic life from impulsive sound produced by activities such as pile driving and seismic exploration is increasing throughout the world, particularly with the increased exploitation of oceans for energy production. At the same time, there are few data that provide insight into the effects of these sounds on fishes. The goal of this study was to provide quantitative data to define the levels of impulsive sound that could result in the onset of barotrauma to fish. A High Intensity Controlled Impedance Fluid filled wave Tube was developed that enabled laboratory simulation of high-energy impulsive sound that were characteristic of aquatic far-field, plane-wave acoustic conditions. The sounds used were based upon the impulsive sounds generated by an impact hammer striking a steel shell pile. Neutrally buoyant juvenile Chinook salmon (Oncorhynchus tshawytscha) were exposed to impulsive sounds and subsequently evaluated for barotrauma injuries. Observed injuries ranged from mild hematomas at the lowest sound exposure levels to organ hemorrhage at the highest sound exposure levels. Frequency of observed injuries were used to compute a biological response weighted index (RWI) to evaluate the physiological impact of injuries at the different exposure levels. As single strike and cumulative sound exposure levels (SELss, SELcum respectively) increased, RWI values increased. Based on the results, tissue damage associated with adverse physiological costs occurred when the RWI was greater than 2. In terms of sound exposure levels a RWI of 2 was achieved for 1920 strikes by 177 dB re 1 µPa2⋅s SELss yielding a SELcum of 210 dB re 1 µPa2⋅s, and for 960 strikes by 180 dB re 1 µPa2⋅s SELss yielding a SELcum of 210 dB re 1 µPa2⋅s. These metrics define thresholds for onset of injury in juvenile Chinook salmon. PMID:22745695
Detection of Mouse Cough Based on Sound Monitoring and Respiratory Airflow Waveforms
Chen, Liyan; Lai, Kefang; Lomask, Joseph Mark; Jiang, Bert; Zhong, Nanshan
2013-01-01
Detection for cough in mice has never yielded clearly audible sounds, so there is still a great deal of debates as to whether mice can cough in response to tussive stimuli. Here we introduce an approach for detection of mouse cough based on sound monitoring and airflow signals. 40 Female BALB/c mice were pretreated with normal saline, codeine, capasazepine or desensitized with capsaicin. Single mouse was put in a plethysmograph, exposed to aerosolized 100 µmol/L capsaicin for 3 min, followed by continuous observation for 3 min. Airflow signals of total 6 min were recorded and analyzed to detect coughs. Simultaneously, mouse cough sounds were sensed by a mini-microphone, monitored manually by an operator. When manual and automatic detection coincided, the cough was positively identified. Sound and sound waveforms were also recorded and filtered for further analysis. Body movements were observed by operator. Manual versus automated counts were compared. Seven types of airflow signals were identified by integrating manual and automated monitoring. Observation of mouse movements and analysis of sound waveforms alone did not produce meaningful data. Mouse cough numbers decreased significantly after all above drugs treatment. The Bland-Altman and consistency analysis between automatic and manual counts was 0.968 and 0.956. The study suggests that the mouse is able to present with cough, which could be detected by sound monitoring and respiratory airflow waveform changes. PMID:23555643
Decibels via Loudness--Distance Properties of a Loudspeaker
ERIC Educational Resources Information Center
Bates, Alan
2014-01-01
Loudness, or sound intensity level, is a human perceived or subjective measurement with units, decibels, based on the response of the human ear to different sound intensities. The response of the human ear at fixed frequency is close to being logarithmic. The experiment reported here investigates the relationship between measured sound intensity…
Environmental Sound Training in Cochlear Implant Users
ERIC Educational Resources Information Center
Shafiro, Valeriy; Sheft, Stanley; Kuvadia, Sejal; Gygi, Brian
2015-01-01
Purpose: The study investigated the effect of a short computer-based environmental sound training regimen on the perception of environmental sounds and speech in experienced cochlear implant (CI) patients. Method: Fourteen CI patients with the average of 5 years of CI experience participated. The protocol consisted of 2 pretests, 1 week apart,…
Measuring the Speed of Sound through Gases Using Nitrocellulose
ERIC Educational Resources Information Center
Molek, Karen Sinclair; Reyes, Karl A.; Burnette, Brandon A.; Stepherson, Jacob R.
2015-01-01
Measuring the heat capacity ratios, ?, of gases either through adiabatic expansion or sound velocity is a well established physical chemistry experiment. The most accurate experiments depend on an exact determination of sound origin, which necessitates the use of lasers or a wave generator, where time zero is based on an electrical trigger. Other…
Physics of thermo-acoustic sound generation
NASA Astrophysics Data System (ADS)
Daschewski, M.; Boehm, R.; Prager, J.; Kreutzbruck, M.; Harrer, A.
2013-09-01
We present a generalized analytical model of thermo-acoustic sound generation based on the analysis of thermally induced energy density fluctuations and their propagation into the adjacent matter. The model provides exact analytical prediction of the sound pressure generated in fluids and solids; consequently, it can be applied to arbitrary thermal power sources such as thermophones, plasma firings, laser beams, and chemical reactions. Unlike existing approaches, our description also includes acoustic near-field effects and sound-field attenuation. Analytical results are compared with measurements of sound pressures generated by thermo-acoustic transducers in air for frequencies up to 1 MHz. The tested transducers consist of titanium and indium tin oxide coatings on quartz glass and polycarbonate substrates. The model reveals that thermo-acoustic efficiency increases linearly with the supplied thermal power and quadratically with thermal excitation frequency. Comparison of the efficiency of our thermo-acoustic transducers with those of piezoelectric-based airborne ultrasound transducers using impulse excitation showed comparable sound pressure values. The present results show that thermo-acoustic transducers can be applied as broadband, non-resonant, high-performance ultrasound sources.
Pulse-echo sound speed estimation using second order speckle statistics
NASA Astrophysics Data System (ADS)
Rosado-Mendez, Ivan M.; Nam, Kibo; Madsen, Ernest L.; Hall, Timothy J.; Zagzebski, James A.
2012-10-01
This work presents a phantom-based evaluation of a method for estimating soft-tissue speeds of sound using pulse-echo data. The method is based on the improvement of image sharpness as the sound speed value assumed during beamforming is systematically matched to the tissue sound speed. The novelty of this work is the quantitative assessment of image sharpness by measuring the resolution cell size from the autocovariance matrix for echo signals from a random distribution of scatterers thus eliminating the need of strong reflectors. Envelope data were obtained from a fatty-tissue mimicking (FTM) phantom (sound speed = 1452 m/s) and a nonfatty-tissue mimicking (NFTM) phantom (1544 m/s) scanned with a linear array transducer on a clinical ultrasound system. Dependence on pulse characteristics was tested by varying the pulse frequency and amplitude. On average, sound speed estimation errors were -0.7% for the FTM phantom and -1.1% for the NFTM phantom. In general, no significant difference was found among errors from different pulse frequencies and amplitudes. The method is currently being optimized for the differentiation of diffuse liver diseases.
2013-01-01
Background In this paper, we developed a novel algorithm to detect the valvular split between the aortic and pulmonary components in the second heart sound which is a valuable medical information. Methods The algorithm is based on the Reassigned smoothed pseudo Wigner–Ville distribution which is a modified time–frequency distribution of the Wigner–Ville distribution. A preprocessing amplitude recovery procedure is carried out on the analysed heart sound to improve the readability of the time–frequency representation. The simulated S2 heart sounds were generated by an overlapping frequency modulated chirp–based model at different valvular split durations. Results Simulated and real heart sounds are processed to highlight the performance of the proposed approach. The algorithm is also validated on real heart sounds of the LGB–IRCM (Laboratoire de Génie biomédical–Institut de recherches cliniques de Montréal) cardiac valve database. The A2–P2 valvular split is accurately detected by processing the obtained RSPWVD representations for both simulated and real data. PMID:23631738
Studies Based on Lunar Global Subsurface Radar Sounding Data Obtained by SELENE (Kaguya)
NASA Astrophysics Data System (ADS)
Kumamoto, A.; Yamaguchi, Y.; Yamaji, A.; Oshigami, S.; Ishiyama, K.; Nakamura, N.; Haruyama, J.; Miyamoto, H.; Nishibori, T.; Tsuchiya, F.; Ohtake, M.
2018-04-01
Several studies based on lunar global subsurface radar sounding data obtained by SELENE/LRS will be reviewed. From the subsurface structures of the buried regolith layers, we can discuss the evolution of tectonic and volcanic processes in the maria.
NASA Astrophysics Data System (ADS)
Elliott, Stephen J.; Cheer, Jordan; Bhan, Lam; Shi, Chuang; Gan, Woon-Seng
2018-04-01
The active control of an incident sound field with an array of secondary sources is a fundamental problem in active control. In this paper the optimal performance of an infinite array of secondary sources in controlling a plane incident sound wave is first considered in free space. An analytic solution for normal incidence plane waves is presented, indicating a clear cut-off frequency for good performance, when the separation distance between the uniformly-spaced sources is equal to a wavelength. The extent of the near field pressure close to the source array is also quantified, since this determines the positions of the error microphones in a practical arrangement. The theory is also extended to oblique incident waves. This result is then compared with numerical simulations of controlling the sound power radiated through an open aperture in a rigid wall, subject to an incident plane wave, using an array of secondary sources in the aperture. In this case the diffraction through the aperture becomes important when its size is compatible with the acoustic wavelength, in which case only a few sources are necessary for good control. When the size of the aperture is large compared to the wavelength, and diffraction is less important but more secondary sources need to be used for good control, the results then become similar to those for the free field problem with an infinite source array.
NASA Astrophysics Data System (ADS)
Maeda, S.; Minami, S.; Okamoto, D.; Obara, T.
2016-09-01
The deflagration-to-detonation transition in a 100 mm square cross-section channel was investigated for a highly reactive stoichiometric hydrogen oxygen mixture at 70 kPa. Obstacles of 5 mm width and 5, 10, and 15 mm heights were equally spaced 60 mm apart at the bottom of the channel. The phenomenon was investigated primarily by time-resolved schlieren visualization from two orthogonal directions using a high-speed video camera. The detonation transition occurred over a remarkably short distance within only three or four repeated obstacles. The global flame speed just before the detonation transition was well below the sound speed of the combustion products and did not reach the sound speed of the initial unreacted gas for tests with an obstacle height of 5 and 10 mm. These results indicate that a detonation transition does not always require global flame acceleration beyond the speed of sound for highly reactive combustible mixtures. A possible mechanism for this detonation initiation was the mixing of the unreacted and reacted gas in the vicinity of the flame front convoluted by the vortex present behind each obstacle, and the formation of a hot spot by the shock wave. The final onset of the detonation originated from the unreacted gas pocket, which was surrounded by the obstacle downstream face and the channel wall.
Kastelein, Ronald A; Hoek, Lean; Gransier, Robin; Rambags, Martijn; Claeys, Naomi
2014-07-01
Safety criteria for underwater low-frequency active sonar sounds produced during naval exercises are needed to protect harbor porpoise hearing. As a first step toward defining criteria, a porpoise was exposed to sequences consisting of series of 1-s, 1-2 kHz sonar down-sweeps without harmonics (as fatiguing noise) at various combinations of average received sound pressure levels (SPLs; 144-179 dB re 1 μPa), exposure durations (1.9-240 min), and duty cycles (5%-100%). Hearing thresholds were determined for a narrow-band frequency-swept sine wave centered at 1.5 kHz before exposure to the fatiguing noise, and at 1-4, 4-8, 8-12, 48, 96, 144, and 1400 min after exposure, to quantify temporary threshold shifts (TTSs) and recovery of hearing. Results show that the inter-pulse interval of the fatiguing noise is an important parameter in determining the magnitude of noise-induced TTS. For the reported range of exposure combinations (duration and SPL), the energy of the exposure (i.e., cumulative sound exposure level; SELcum) can be used to predict the induced TTS, if the inter-pulse interval is known. Exposures with equal SELcum but with different inter-pulse intervals do not result in the same induced TTS.
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
2010-09-01
BEETIT Project: Penn State is designing a freezer that substitutes the use of sound waves and environmentally benign refrigerant for synthetic refrigerants found in conventional freezers. Called a thermoacoustic chiller, the technology is based on the fact that the pressure oscillations in a sound wave result in temperature changes. Areas of higher pressure raise temperatures and areas of low pressure decrease temperatures. By carefully arranging a series of heat exchangers in a sound field, the chiller is able to isolate the hot and cold regions of the sound waves. Penn State’s chiller uses helium gas to replace synthetic refrigerants. Becausemore » helium does not burn, explode or combine with other chemicals, it is an environmentally-friendly alternative to other polluting refrigerants. Penn State is working to apply this technology on a large scale.« less
Enabling People Who Are Blind to Experience Science Inquiry Learning through Sound-Based Mediation
ERIC Educational Resources Information Center
Levy, S. T.; Lahav, O.
2012-01-01
This paper addresses a central need among people who are blind, access to inquiry-based science learning materials, which are addressed by few other learning environments that use assistive technologies. In this study, we investigated ways in which learning environments based on sound mediation can support science learning by blind people. We used…
ERIC Educational Resources Information Center
Oliveira, Carla; Lousada, Marisa; Jesus, Luis M. T.
2015-01-01
Children with speech sound disorders (SSD) represent a large number of speech and language therapists' caseloads. The intervention with children who have SSD can involve different therapy approaches, and these may be articulatory or phonologically based. Some international studies reveal a widespread application of articulatory based approaches in…
Nonlinear rescaling of control values simplifies fuzzy control
NASA Technical Reports Server (NTRS)
Vanlangingham, H.; Tsoukkas, A.; Kreinovich, V.; Quintana, C.
1993-01-01
Traditional control theory is well-developed mainly for linear control situations. In non-linear cases there is no general method of generating a good control, so we have to rely on the ability of the experts (operators) to control them. If we want to automate their control, we must acquire their knowledge and translate it into a precise control strategy. The experts' knowledge is usually represented in non-numeric terms, namely, in terms of uncertain statements of the type 'if the obstacle is straight ahead, the distance to it is small, and the velocity of the car is medium, press the brakes hard'. Fuzzy control is a methodology that translates such statements into precise formulas for control. The necessary first step of this strategy consists of assigning membership functions to all the terms that the expert uses in his rules (in our sample phrase these words are 'small', 'medium', and 'hard'). The appropriate choice of a membership function can drastically improve the quality of a fuzzy control. In the simplest cases, we can take the functions whose domains have equally spaced endpoints. Because of that, many software packages for fuzzy control are based on this choice of membership functions. This choice is not very efficient in more complicated cases. Therefore, methods have been developed that use neural networks or generic algorithms to 'tune' membership functions. But this tuning takes lots of time (for example, several thousands iterations are typical for neural networks). In some cases there are evident physical reasons why equally space domains do not work: e.g., if the control variable u is always positive (i.e., if we control temperature in a reactor), then negative values (that are generated by equal spacing) simply make no sense. In this case it sounds reasonable to choose another scale u' = f(u) to represent u, so that equal spacing will work fine for u'. In the present paper we formulate the problem of finding the best rescaling function, solve this problem, and show (on a real-life example) that after an optimal rescaling, the un-tuned fuzzy control can be as good as the best state-of-art traditional non-linear controls.
Central pattern generator for vocalization: is there a vertebrate morphotype?
Bass, Andrew H
2014-10-01
Animals that generate acoustic signals for social communication are faced with two essential tasks: generate a temporally precise signal and inform the auditory system about the occurrence of one's own sonic signal. Recent studies of sound producing fishes delineate a hindbrain network comprised of anatomically distinct compartments coding equally distinct neurophysiological properties that allow an organism to meet these behavioral demands. A set of neural characters comprising a vocal-sonic central pattern generator (CPG) morphotype is proposed for fishes and tetrapods that shares evolutionary developmental origins with pectoral appendage motor systems. Copyright © 2014 Elsevier Ltd. All rights reserved.
Central pattern generator for vocalization: Is there a vertebrate morphotype?
Bass, Andrew H.
2014-01-01
Animals that generate acoustic signals for social communication are faced with two essential tasks: generate a temporally precise signal and inform the auditory system about the occurrence of one’s own sonic signal. Recent studies of sound producing fishes delineate a hindbrain network comprised of anatomically distinct compartments coding equally distinct neurophysiological properties that allow an organism to meet these behavioral demands. A set of neural characters comprising a vocal-sonic central pattern generator (CPG) morphotype is proposed for fishes and tetrapods that shares evolutionary developmental origins with pectoral appendage motor systems. PMID:25050813
1968-01-01
This is a cutaway view of the Saturn V first stage, known as the S-IC, detailing the five F-1 engines and fuel cells. The S-IC stage is 138 feet long and 33 feet in diameter, producing more than 7,500,000 pounds of thrust through the five F-1 engines that are powered by liquid oxygen and kerosene. Four of the engines are mounted on an outer ring and gimbal for control purposes. The fifth engine is rigidly mounted in the center. When ignited, the roar produced by the five engines equals the sound of 8,000,000 hi-fi sets.
Topological phononic states of underwater sound based on coupled ring resonators
DOE Office of Scientific and Technical Information (OSTI.GOV)
He, Cheng; Li, Zheng; Ni, Xu
We report a design of topological phononic states for underwater sound using arrays of acoustic coupled ring resonators. In each individual ring resonator, two degenerate acoustic modes, corresponding to clockwise and counter-clockwise propagation, are treated as opposite pseudospins. The gapless edge states arise in the bandgap resulting in protected pseudospin-dependent sound transportation, which is a phononic analogue of the quantum spin Hall effect. We also investigate the robustness of the topological sound state, suggesting that the observed pseudospin-dependent sound transportation remains unless the introduced defects facilitate coupling between the clockwise and counter-clockwise modes (in other words, the original mode degeneracymore » is broken). The topological engineering of sound transportation will certainly promise unique design for next generation of acoustic devices in sound guiding and switching, especially for underwater acoustic devices.« less
Jet-noise reduction through liquid-base foam injection.
NASA Technical Reports Server (NTRS)
Manson, L.; Burge, H. L.
1971-01-01
An experimental investigation has been made of the sound-absorbing properties of liquid-base foams and of their ability to reduce jet noise. Protein, detergent, and polymer foaming agents were used in water solutions. A method of foam generation was developed to permit systematic variation of the foam density. The investigation included measurements of sound-absorption coefficents for both plane normal incidence waves and diffuse sound fields. The intrinsic acoustic properties of foam, e.g., the characteristic impedance and the propagation constant, were also determined. The sound emitted by a 1-in.-diam cold nitrogen jet was measured for subsonic (300 m/sec) and supersonic (422 m/sec) jets, with and without foam injection. Noise reductions up to 10 PNdB were measured.
Digital Sound Encryption with Logistic Map and Number Theoretic Transform
NASA Astrophysics Data System (ADS)
Satria, Yudi; Gabe Rizky, P. H.; Suryadi, MT
2018-03-01
Digital sound security has limits on encrypting in Frequency Domain. Number Theoretic Transform based on field (GF 2521 – 1) improve and solve that problem. The algorithm for this sound encryption is based on combination of Chaos function and Number Theoretic Transform. The Chaos function that used in this paper is Logistic Map. The trials and the simulations are conducted by using 5 different digital sound files data tester in Wave File Extension Format and simulated at least 100 times each. The key stream resulted is random with verified by 15 NIST’s randomness test. The key space formed is very big which more than 10469. The processing speed of algorithm for encryption is slightly affected by Number Theoretic Transform.
Park, Steve; Guan, Xiying; Kim, Youngwan; Creighton, Francis Pete X; Wei, Eric; Kymissis, Ioannis John; Nakajima, Hideko Heidi; Olson, Elizabeth S
2018-01-01
We report the fabrication and characterization of a prototype polyvinylidene fluoride polymer-based implantable microphone for detecting sound inside gerbil and human cochleae. With the current configuration and amplification, the signal-to-noise ratios were sufficiently high for normally occurring sound pressures and frequencies (ear canal pressures >50-60 dB SPL and 0.1-10 kHz), though 10 to 20 dB poorer than for some hearing aid microphones. These results demonstrate the feasibility of the prototype devices as implantable microphones for the development of totally implantable cochlear implants. For patients, this will improve sound reception by utilizing the outer ear and will improve the use of cochlear implants.
Guan, Xiying; Kim, Youngwan; Creighton, Francis (Pete) X.; Wei, Eric; Kymissis, Ioannis(John); Nakajima, Hideko Heidi; Olson, Elizabeth S.
2018-01-01
We report the fabrication and characterization of a prototype polyvinylidene fluoride polymer-based implantable microphone for detecting sound inside gerbil and human cochleae. With the current configuration and amplification, the signal-to-noise ratios were sufficiently high for normally occurring sound pressures and frequencies (ear canal pressures >50–60 dB SPL and 0.1–10 kHz), though 10 to 20 dB poorer than for some hearing aid microphones. These results demonstrate the feasibility of the prototype devices as implantable microphones for the development of totally implantable cochlear implants. For patients, this will improve sound reception by utilizing the outer ear and will improve the use of cochlear implants. PMID:29732950
The Effect of Room Acoustics on the Sleep Quality of Healthy Sleepers
Fietze, Ingo; Barthe, Charlotte; Hölzl, Matthias; Glos, Martin; Zimmermann, Sandra; Bauer-Diefenbach, Ralf; Penzel, Thomas
2016-01-01
Introduction: Noise is one of the factors that can seriously disturb sleep, and sound volume is an important factor in this context. One strategy involves avoiding exposure to sounds in the night, while entail the minimization of background noise in a bedroom. The goal of this study was to investigate the effect of systematic sound attenuation on nocturnal sleep by influencing sound volume and reverberation within the context of room acoustics. Materials and Methods: On this basis, we designed a randomized, controlled crossover trial investigating 24 healthy sleepers (15 men and 9 women, aged 24.9 ± 4.1 years) with a body mass index (BMI) of 21.9 ± 1.6 kg/m2. Each participant slept for three consecutive nights at three different locations: (a) at our sleep lab, (b) at the participant's home, and (c) at an acoustically isolated room. In addition to conduct of polysomnography (PSG), subjective sleep quality and nocturnal noise level were measured at each location. We likewise measured room temperature and relative humidity. Results: Under conditions of equal sleep efficiency, a significant increase in deep sleep, by 16–34 min, was determined in an acoustically isolated room in comparison to the two other sleep locations. Fewer arousal events and an increase in rapid eye movement (REM) latency became evident in an acoustically isolated environment. Sleep in a domestic environment was subjectively better than sleep under the two test conditions. Discussion: For healthy sleepers, room acoustics influence the microstructure of sleep, without subjective morning benefit. Reduction of noise level and of reverberation leads to an increase in the amount of deep sleep and to reduction of nocturnal arousal events, which is especially important for poor sleepers. PMID:27762252
The effect of room acoustics on the sleep quality of healthy sleepers.
Fietze, Ingo; Barthe, Charlotte; Hölzl, Matthias; Glos, Martin; Zimmermann, Sandra; Bauer-Diefenbach, Ralf; Penzel, Thomas
2016-01-01
Noise is one of the factors that can seriously disturb sleep, and sound volume is an important factor in this context. One strategy involves avoiding exposure to sounds in the night, while entail the minimization of background noise in a bedroom. The goal of this study was to investigate the effect of systematic sound attenuation on nocturnal sleep by influencing sound volume and reverberation within the context of room acoustics. On this basis, we designed a randomized, controlled crossover trial investigating 24 healthy sleepers (15 men and 9 women, aged 24.9 ± 4.1 years) with a body mass index (BMI) of 21.9 ± 1.6 kg/m2. Each participant slept for three consecutive nights at three different locations: (a) at our sleep lab, (b) at the participant's home, and (c) at an acoustically isolated room. In addition to conduct of polysomnography (PSG), subjective sleep quality and nocturnal noise level were measured at each location. We likewise measured room temperature and relative humidity. Under conditions of equal sleep efficiency, a significant increase in deep sleep, by 16-34 min, was determined in an acoustically isolated room in comparison to the two other sleep locations. Fewer arousal events and an increase in rapid eye movement (REM) latency became evident in an acoustically isolated environment. Sleep in a domestic environment was subjectively better than sleep under the two test conditions. For healthy sleepers, room acoustics influence the microstructure of sleep, without subjective morning benefit. Reduction of noise level and of reverberation leads to an increase in the amount of deep sleep and to reduction of nocturnal arousal events, which is especially important for poor sleepers.
Relative size of auditory pathways in symmetrically and asymmetrically eared owls.
Gutiérrez-Ibáñez, Cristián; Iwaniuk, Andrew N; Wylie, Douglas R
2011-01-01
Owls are highly efficient predators with a specialized auditory system designed to aid in the localization of prey. One of the most unique anatomical features of the owl auditory system is the evolution of vertically asymmetrical ears in some species, which improves their ability to localize the elevational component of a sound stimulus. In the asymmetrically eared barn owl, interaural time differences (ITD) are used to localize sounds in azimuth, whereas interaural level differences (ILD) are used to localize sounds in elevation. These two features are processed independently in two separate neural pathways that converge in the external nucleus of the inferior colliculus to form an auditory map of space. Here, we present a comparison of the relative volume of 11 auditory nuclei in both the ITD and the ILD pathways of 8 species of symmetrically and asymmetrically eared owls in order to investigate evolutionary changes in the auditory pathways in relation to ear asymmetry. Overall, our results indicate that asymmetrically eared owls have much larger auditory nuclei than owls with symmetrical ears. In asymmetrically eared owls we found that both the ITD and ILD pathways are equally enlarged, and other auditory nuclei, not directly involved in binaural comparisons, are also enlarged. We suggest that the hypertrophy of auditory nuclei in asymmetrically eared owls likely reflects both an improved ability to precisely locate sounds in space and an expansion of the hearing range. Additionally, our results suggest that the hypertrophy of nuclei that compute space may have preceded that of the expansion of the hearing range and evolutionary changes in the size of the auditory system occurred independently of phylogeny. Copyright © 2011 S. Karger AG, Basel.
Siveke, Ida; Leibold, Christian; Grothe, Benedikt
2007-11-01
We are regularly exposed to several concurrent sounds, producing a mixture of binaural cues. The neuronal mechanisms underlying the localization of concurrent sounds are not well understood. The major binaural cues for localizing low-frequency sounds in the horizontal plane are interaural time differences (ITDs). Auditory brain stem neurons encode ITDs by firing maximally in response to "favorable" ITDs and weakly or not at all in response to "unfavorable" ITDs. We recorded from ITD-sensitive neurons in the dorsal nucleus of the lateral lemniscus (DNLL) while presenting pure tones at different ITDs embedded in noise. We found that increasing levels of concurrent white noise suppressed the maximal response rate to tones with favorable ITDs and slightly enhanced the response rate to tones with unfavorable ITDs. Nevertheless, most of the neurons maintained ITD sensitivity to tones even for noise intensities equal to that of the tone. Using concurrent noise with a spectral composition in which the neuron's excitatory frequencies are omitted reduced the maximal response similar to that obtained with concurrent white noise. This finding indicates that the decrease of the maximal rate is mediated by suppressive cross-frequency interactions, which we also observed during monaural stimulation with additional white noise. In contrast, the enhancement of the firing rate to tones at unfavorable ITD might be due to early binaural interactions (e.g., at the level of the superior olive). A simple simulation corroborates this interpretation. Taken together, these findings suggest that the spectral composition of a concurrent sound strongly influences the spatial processing of ITD-sensitive DNLL neurons.
Rising tones and rustling noises: Metaphors in gestural depictions of sounds
Scurto, Hugo; Françoise, Jules; Bevilacqua, Frédéric; Houix, Olivier; Susini, Patrick
2017-01-01
Communicating an auditory experience with words is a difficult task and, in consequence, people often rely on imitative non-verbal vocalizations and gestures. This work explored the combination of such vocalizations and gestures to communicate auditory sensations and representations elicited by non-vocal everyday sounds. Whereas our previous studies have analyzed vocal imitations, the present research focused on gestural depictions of sounds. To this end, two studies investigated the combination of gestures and non-verbal vocalizations. A first, observational study examined a set of vocal and gestural imitations of recordings of sounds representative of a typical everyday environment (ecological sounds) with manual annotations. A second, experimental study used non-ecological sounds whose parameters had been specifically designed to elicit the behaviors highlighted in the observational study, and used quantitative measures and inferential statistics. The results showed that these depicting gestures are based on systematic analogies between a referent sound, as interpreted by a receiver, and the visual aspects of the gestures: auditory-visual metaphors. The results also suggested a different role for vocalizations and gestures. Whereas the vocalizations reproduce all features of the referent sounds as faithfully as vocally possible, the gestures focus on one salient feature with metaphors based on auditory-visual correspondences. Both studies highlighted two metaphors consistently shared across participants: the spatial metaphor of pitch (mapping different pitches to different positions on the vertical dimension), and the rustling metaphor of random fluctuations (rapidly shaking of hands and fingers). We interpret these metaphors as the result of two kinds of representations elicited by sounds: auditory sensations (pitch and loudness) mapped to spatial position, and causal representations of the sound sources (e.g. rain drops, rustling leaves) pantomimed and embodied by the participants’ gestures. PMID:28750071
Lung and Heart Sounds Analysis: State-of-the-Art and Future Trends.
Padilla-Ortiz, Ana L; Ibarra, David
2018-01-01
Lung sounds, which include all sounds that are produced during the mechanism of respiration, may be classified into normal breath sounds and adventitious sounds. Normal breath sounds occur when no respiratory problems exist, whereas adventitious lung sounds (wheeze, rhonchi, crackle, etc.) are usually associated with certain pulmonary pathologies. Heart and lung sounds that are heard using a stethoscope are the result of mechanical interactions that indicate operation of cardiac and respiratory systems, respectively. In this article, we review the research conducted during the last six years on lung and heart sounds, instrumentation and data sources (sensors and databases), technological advances, and perspectives in processing and data analysis. Our review suggests that chronic obstructive pulmonary disease (COPD) and asthma are the most common respiratory diseases reported on in the literature; related diseases that are less analyzed include chronic bronchitis, idiopathic pulmonary fibrosis, congestive heart failure, and parenchymal pathology. Some new findings regarding the methodologies associated with advances in the electronic stethoscope have been presented for the auscultatory heart sound signaling process, including analysis and clarification of resulting sounds to create a diagnosis based on a quantifiable medical assessment. The availability of automatic interpretation of high precision of heart and lung sounds opens interesting possibilities for cardiovascular diagnosis as well as potential for intelligent diagnosis of heart and lung diseases.
NASA Astrophysics Data System (ADS)
Sato, Shin-Ichi; Prodi, Nicola; Sakai, Hiroyuki
2004-05-01
To clarify the relationship of the sound fields between the stage and the orchestra pit, we conducted acoustical measurements in a typical historical opera house, the Teatro Comunale of Ferrara, Italy. Orthogonal factors based on the theory of subjective preference and other related factors were analyzed. First, the sound fields for a singer on the stage in relation to the musicians in the pit were analyzed. And then, the sound fields for performers in the pit in relation to the singers on the stage were considered. Because physical factors vary depending on the location of the sound source, performers can move on the stage or in the pit to find the preferred sound field.
[A focused sound field measurement system by LabVIEW].
Jiang, Zhan; Bai, Jingfeng; Yu, Ying
2014-05-01
In this paper, according to the requirement of the focused sound field measurement, a focused sound field measurement system was established based on the LabVIEW virtual instrument platform. The system can automatically search the focus position of the sound field, and adjust the scanning path according to the size of the focal region. Three-dimensional sound field scanning time reduced from 888 hours in uniform step to 9.25 hours in variable step. The efficiency of the focused sound field measurement was improved. There is a certain deviation between measurement results and theoretical calculation results. Focal plane--6 dB width difference rate was 3.691%, the beam axis--6 dB length differences rate was 12.937%.
Makeyev, Oleksandr; Sazonov, Edward; Schuckers, Stephanie; Lopez-Meyer, Paulo; Melanson, Ed; Neuman, Michael
2007-01-01
In this paper we propose a sound recognition technique based on the limited receptive area (LIRA) neural classifier and continuous wavelet transform (CWT). LIRA neural classifier was developed as a multipurpose image recognition system. Previous tests of LIRA demonstrated good results in different image recognition tasks including: handwritten digit recognition, face recognition, metal surface texture recognition, and micro work piece shape recognition. We propose a sound recognition technique where scalograms of sound instances serve as inputs of the LIRA neural classifier. The methodology was tested in recognition of swallowing sounds. Swallowing sound recognition may be employed in systems for automated swallowing assessment and diagnosis of swallowing disorders. The experimental results suggest high efficiency and reliability of the proposed approach.
Chestnut, Eleanor K; Markman, Ellen M
2018-06-28
Although "Girls are as good as boys at math" explicitly expresses equality, we predict it could nevertheless suggest that boys have more raw talent. In statements with this subject-complement structure, the item in the complement position serves as the reference point and is thus considered more typical and prominent. This explains why "Tents are like houses," for instance, sounds better than "Houses are like tents"-people generally think of houses as more typical. For domains about ability, the reference point should be the item that is typically more skilled. We further propose that the reference point should be naturally more skilled. In two experiments, we presented adults with summaries of actual scientific evidence for gender equality in math (Experiment 1) or verbal ability (Experiment 2), but we manipulated whether the reference point in the statements of equality in the summaries (e.g., "Boys' verbal ability is as good as girls'") was girls or boys. As predicted, adults attributed more natural ability to each gender when it was in the complement rather than subject position. Yet, in Experiment 3, we found that when explicitly asked, participants judged that such sentences were not biased in favor of either gender, indicating that subject-complement statements must be transmitting this bias in a subtle way. Thus, statements such as "Girls are as good as boys at math" can actually backfire and perpetuate gender stereotypes about natural ability. © 2018 Cognitive Science Society, Inc.
New Stethoscope With Extensible Diaphragm.
Takashina, Tsunekazu; Shimizu, Masashi; Muratake, Torakazu; Mayuzumi, Syuichi
2016-08-25
This study compared the diagnostic efficacy of the common suspended diaphragm stethoscope (SDS) with a new extensible diaphragm stethoscope (EDS) for low-frequency heart sounds. The EDS was developed by using an ethylene propylene diene monomer diaphragm. The results showed that the EDS enhanced both the volume and quality of low-frequency heart sounds, and improved the ability of examiners to auscultate such heart sounds. Based on the results of the sound analysis, the EDS is more efficient than the SDS. (Circ J 2016; 80: 2047-2049).
2017-02-01
difference from the climate -based METCM. The Tv changes are shown in Fig. 5, but given the smaller relative changes only the ±2 SD curves are presented...planning and in field tests when sounding data are not available. However, the use of climate mean profiles may lead to wide differences from actual...individual atmospheric profiles. This brief report investigates the variation of a series of soundings as compared to climate mean soundings and
Feasibility of making sound power measurements in the NASA Langley V/STOL tunnel test section
NASA Technical Reports Server (NTRS)
Brooks, T. F.; Scheiman, J.; Silcox, R. J.
1976-01-01
Based on exploratory acoustic measurements in Langley's V/STOL wind tunnel, recommendations are made on the methodology for making sound power measurements of aircraft components in the closed tunnel test section. During airflow, tunnel self-noise and microphone flow-induced noise place restrictions on the amplitude and spectrum of the sound source to be measured. Models of aircraft components with high sound level sources, such as thrust engines and powered lift systems, seem likely candidates for acoustic testing.
On the sound insulation of acoustic metasurface using a sub-structuring approach
NASA Astrophysics Data System (ADS)
Yu, Xiang; Lu, Zhenbo; Cheng, Li; Cui, Fangsen
2017-08-01
The feasibility of using an acoustic metasurface (AMS) with acoustic stop-band property to realize sound insulation with ventilation function is investigated. An efficient numerical approach is proposed to evaluate its sound insulation performance. The AMS is excited by a reverberant sound source and the standardized sound reduction index (SRI) is numerically investigated. To facilitate the modeling, the coupling between the AMS and the adjacent acoustic fields is formulated using a sub-structuring approach. A modal based formulation is applied to both the source and receiving room, enabling an efficient calculation in the frequency range from 125 Hz to 2000 Hz. The sound pressures and the velocities at the interface are matched by using a transfer function relation based on ;patches;. For illustration purposes, numerical examples are investigated using the proposed approach. The unit cell constituting the AMS is constructed in the shape of a thin acoustic chamber with tailored inner structures, whose stop-band property is numerically analyzed and experimentally demonstrated. The AMS is shown to provide effective sound insulation of over 30 dB in the stop-band frequencies from 600 to 1600 Hz. It is also shown that the proposed approach has the potential to be applied to a broad range of AMS studies and optimization problems.
NASA Astrophysics Data System (ADS)
Ma, Fuyin; Wu, Jiu Hui; Huang, Meng
2015-09-01
In order to overcome the influence of the structural resonance on the continuous structures and obtain a lightweight thin-layer structure which can effectively isolate the low-frequency noises, an elastic membrane structure was proposed. In the low-frequency range below 500 Hz, the sound transmission loss (STL) of this membrane type structure is greatly higher than that of the current sound insulation material EVA (ethylene-vinyl acetate copo) of vehicle, so it is possible to replace the EVA by the membrane-type metamaterial structure in practice engineering. Based on the band structure, modal shapes, as well as the sound transmission simulation, the sound insulation mechanism of the designed membrane-type acoustic metamaterials was analyzed from a new perspective, which had been validated experimentally. It is suggested that in the frequency range above 200 Hz for this membrane-mass type structure, the sound insulation effect was principally not due to the low-level locally resonant mode of the mass block, but the continuous vertical resonant modes of the localized membrane. So based on such a physical property, a resonant modal group theory is initially proposed in this paper. In addition, the sound insulation mechanism of the membrane-type structure and thin plate structure were combined by the membrane/plate resonant theory.
Evaluation of a low-cost 3D sound system for immersive virtual reality training systems.
Doerr, Kai-Uwe; Rademacher, Holger; Huesgen, Silke; Kubbat, Wolfgang
2007-01-01
Since Head Mounted Displays (HMD), datagloves, tracking systems, and powerful computer graphics resources are nowadays in an affordable price range, the usage of PC-based "Virtual Training Systems" becomes very attractive. However, due to the limited field of view of HMD devices, additional modalities have to be provided to benefit from 3D environments. A 3D sound simulation can improve the capabilities of VR systems dramatically. Unfortunately, realistic 3D sound simulations are expensive and demand a tremendous amount of computational power to calculate reverberation, occlusion, and obstruction effects. To use 3D sound in a PC-based training system as a way to direct and guide trainees to observe specific events in 3D space, a cheaper alternative has to be provided, so that a broader range of applications can take advantage of this modality. To address this issue, we focus in this paper on the evaluation of a low-cost 3D sound simulation that is capable of providing traceable 3D sound events. We describe our experimental system setup using conventional stereo headsets in combination with a tracked HMD device and present our results with regard to precision, speed, and used signal types for localizing simulated sound events in a virtual training environment.
Scattering Cross Section of Sound Waves by the Modal Element Method
NASA Technical Reports Server (NTRS)
Baumeister, Kenneth J.; Kreider, Kevin L.
1994-01-01
#he modal element method has been employed to determine the scattered field from a plane acoustic wave impinging on a two dimensional body. In the modal element method, the scattering body is represented by finite elements, which are coupled to an eigenfunction expansion representing the acoustic pressure in the infinite computational domain surrounding the body. The present paper extends the previous work by developing the algorithm necessary to calculate the acoustics scattering cross section by the modal element method. The scattering cross section is the acoustical equivalent to the Radar Cross Section (RCS) in electromagnetic theory. Since the scattering cross section is evaluated at infinite distance from the body, an asymptotic approximation is used in conjunction with the standard modal element method. For validation, the scattering cross section of the rigid circular cylinder is computed for the frequency range 0.1 is less than or equal to ka is less than or equal to 100. Results show excellent agreement with the analytic solution.
Will Biomedical Enhancements Undermine Solidarity, Responsibility, Equality and Autonomy?
Lev, Ori
2009-01-01
Prominent thinkers such as Jurgen Habermas and Michael Sandel are warning that biomedical enhancements will undermine fundamental political values. Yet, whether biomedical enhancements will undermine such values depends on how biomedical enhancements will function, how they will be administered and to whom. Since only few enhancements are obtainable, it is difficult to tell whether their predictions are sound. Nevertheless, such warnings are extremely valuable. As a society we must, at the very least, be aware of developments that could have harmful consequences. Indeed, if important values would be jeopardized, we should take appropriate measures to protect them. This paper focuses on four central values: solidarity, personal responsibility, equality and autonomy. It delineates the conditions under which biomedical enhancements would undermine these values. It also details the circumstances under which these values would be unaffected by enhancements as well as those under which they would be promoted. Specifying these conditions is valuable; it would enable society to prepare appropriate ethical guidelines and policy responses in advance. PMID:20002073
NASA Astrophysics Data System (ADS)
Sondkar, Pravin B.
The severity of combined aerodynamics and power transmission response in high-speed, high power density systems such as a rotorcraft is still a major cause of annoyance in spite of recent advancement in passive, semi-active and active control. With further increase in the capacity and power of this class of machinery systems, the acoustic noise levels are expected to increase even more. To achieve further improvements in sound quality, a more refined understanding of the factors and attributes controlling human perception is needed. In the case of rotorcraft systems, the perceived quality of the interior sound field is a major determining factor of passenger comfort. Traditionally, this sound quality factor is determined by measuring the response of a chosen set of juries who are asked to compare their qualitative reactions to two or more sounds based on their subjective impressions. This type of testing is very time-consuming, costly, often inconsistent, and not useful for practical design purposes. Furthermore, there is no known universal model for sound quality. The primary aim of this research is to achieve significant improvements in quantifying the sound quality of combined aerodynamic and power transmission response in high-speed, high power density machinery systems such as a rotorcraft by applying relevant objective measures related to the spectral characteristics of the sound field. Two models have been proposed in this dissertation research. First, a classical multivariate regression analysis model based on currently known sound quality metrics as well some new metrics derived in this study is presented. Even though the analysis resulted in the best possible multivariate model as a measure of the acoustic noise quality, it lacks incorporation of human judgment mechanism. The regression model can change depending on specific application, nature of the sounds and types of juries used in the study. Also, it predicts only the averaged preference scores and does not explain why two jury members differ in their judgment. To address the above shortcoming of applying regression analysis, a new human judgment model is proposed to further improve the ability to predict the degree of subjective annoyance. The human judgment model involves extraction of subjective attributes and their values using a proposed artificial jury processor. In this approach, a set of ear transfer functions are employed to compute the characteristics of sound pressure waves as perceived subjectively by human. The resulting basilar membrane displacement data from this proposed model is then applied to analyze the attribute values. Using this proposed human judgment model, the human judgment mechanism, which is highly sophisticated, will be examined. Since the human judgment model is essentially based on jury attributes that are not expected to change significantly with application or nature of the sound field, it gives a more common basis to evaluate sound quality. This model also attempts to explain the inter-juror differences in opinion, which is critical in understanding the variability in human response.
Excimer laser coronary angioplasty: relative risk analysis of clinical results
NASA Astrophysics Data System (ADS)
Bittl, John A.
1992-08-01
Reports of successful use of excimer laser coronary angioplasty for complex coronary artery disease abound, yet firm indications for its use have not been defined. We attempted to treat 858 coronary stenoses in 764 consecutive patients (mean age 61 years; range 32 - 91 years; 75% men; 76% with Class III or IV angina) with excimer laser angioplasty at 308 nm. Successful treatment was achieved in 86% of patients, as indicated by
NASA Astrophysics Data System (ADS)
Kapranov, B. I.; Mashanov, A. P.
2017-04-01
This paper presents the results of research and describes the apparatus for measuring the acoustic characteristics of bulk materials. Ultrasound, it has passed through a layer of bulk material, is further passes through an air gap. The presence of air gap prevents from measuring tract mechanical contacts, but complicates the measurement technology Studies were conducted on the example of measuring the acoustic characteristics of the widely used perlite-based sound-proofing material.
Re-Sonification of Objects, Events, and Environments
NASA Astrophysics Data System (ADS)
Fink, Alex M.
Digital sound synthesis allows the creation of a great variety of sounds. Focusing on interesting or ecologically valid sounds for music, simulation, aesthetics, or other purposes limits the otherwise vast digital audio palette. Tools for creating such sounds vary from arbitrary methods of altering recordings to precise simulations of vibrating objects. In this work, methods of sound synthesis by re-sonification are considered. Re-sonification, herein, refers to the general process of analyzing, possibly transforming, and resynthesizing or reusing recorded sounds in meaningful ways, to convey information. Applied to soundscapes, re-sonification is presented as a means of conveying activity within an environment. Applied to the sounds of objects, this work examines modeling the perception of objects as well as their physical properties and the ability to simulate interactive events with such objects. To create soundscapes to re-sonify geographic environments, a method of automated soundscape design is presented. Using recorded sounds that are classified based on acoustic, social, semantic, and geographic information, this method produces stochastically generated soundscapes to re-sonify selected geographic areas. Drawing on prior knowledge, local sounds and those deemed similar comprise a locale's soundscape. In the context of re-sonifying events, this work examines processes for modeling and estimating the excitations of sounding objects. These include plucking, striking, rubbing, and any interaction that imparts energy into a system, affecting the resultant sound. A method of estimating a linear system's input, constrained to a signal-subspace, is presented and applied toward improving the estimation of percussive excitations for re-sonification. To work toward robust recording-based modeling and re-sonification of objects, new implementations of banded waveguide (BWG) models are proposed for object modeling and sound synthesis. Previous implementations of BWGs use arbitrary model parameters and may produce a range of simulations that do not match digital waveguide or modal models of the same design. Subject to linear excitations, some models proposed here behave identically to other equivalently designed physical models. Under nonlinear interactions, such as bowing, many of the proposed implementations exhibit improvements in the attack characteristics of synthesized sounds.
Evolution of Signaling in a Multi-Robot System: Categorization and Communication
NASA Astrophysics Data System (ADS)
Ampatzis, Christos; Tuci, Elio; Trianni, Vito; Dorigo, Marco
We use Evolutionary Robotics to design robot controllers in which decision-making mechanisms to switch from solitary to social behavior are integrated with the mechanisms that underpin the sensory-motor repertoire of the robots. In particular, we study the evolution of behavioral and communicative skills in a categorization task. The individual decision-making structures are based on the integration over time of sensory information. The mechanisms for switching from solitary to social behavior and the ways in which the robots can affect each other's behavior are not predetermined by the experimenter, but are aspects of our model designed by artificial evolution. Our results show that evolved robots manage to cooperate and collectively discriminate between different environments by developing a simple communication protocol based on sound signaling. Communication emerges in the absence of explicit selective pressure coded in the fitness function. The evolution of communication is neither trivial nor obvious; for a meaningful signaling system to evolve, evolution must produce both appropriate signals and appropriate reactions to signals. The use of communication proves to be adaptive for the group, even if, in principle, non-cooperating robots can be equally successful with cooperating robots.
THE USE OF ARCHITECTURAL ACOUSTICAL MATERIALS, THEORY AND PRACTICE. SECOND EDITION.
ERIC Educational Resources Information Center
Acoustical Materials Association, New York, NY.
THIS DISCUSSION OF THE BASIC FUNCTION OF ACOUSTICAL MATERIALS--THE CONTROL OF SOUND BY SOUND ABSORPTION--IS BASED ON THE WAVE AND ENERGY PROPERTIES OF SOUND. IT IS STATED THAT, IN GENERAL, A MUCH LARGER VOLUME OF ACOUSTICAL MATERIALS IS NEEDED TO REMOVE DISTRACTING NOISE FROM CLASSROOMS AND OFFICES, FOR EXAMPLE, THAN FROM AUDITORIUMS, WHERE A…
Why Should We Care about Sound Symbolism in EFL Learning?: Two Pilot Studies
ERIC Educational Resources Information Center
Roig-Marín, Amanda
2017-01-01
Given the importance of the phonological and lexical components of the language in L2 learning, this article discusses an innovative, holistic approach to learning these two components of the language based on the existence of "sound symbolism"--the interrelation between sound and meaning--in English. In particular, it describes how and…
Automatic Sound Generation for Spherical Objects Hitting Straight Beams Based on Physical Models.
ERIC Educational Resources Information Center
Rauterberg, M.; And Others
Sounds are the result of one or several interactions between one or several objects at a certain place and in a certain environment; the attributes of every interaction influence the generated sound. The following factors influence users in human/computer interaction: the organization of the learning environment, the content of the learning tasks,…
Motor-Based Treatment with and without Ultrasound Feedback for Residual Speech-Sound Errors
ERIC Educational Resources Information Center
Preston, Jonathan L.; Leece, Megan C.; Maas, Edwin
2017-01-01
Background: There is a need to develop effective interventions and to compare the efficacy of different interventions for children with residual speech-sound errors (RSSEs). Rhotics (the r-family of sounds) are frequently in error American English-speaking children with RSSEs and are commonly targeted in treatment. One treatment approach involves…
Shaping reverberating sound fields with an actively tunable metasurface.
Ma, Guancong; Fan, Xiying; Sheng, Ping; Fink, Mathias
2018-06-26
A reverberating environment is a common complex medium for airborne sound, with familiar examples such as music halls and lecture theaters. The complexity of reverberating sound fields has hindered their meaningful control. Here, by combining acoustic metasurface and adaptive wavefield shaping, we demonstrate the versatile control of reverberating sound fields in a room. This is achieved through the design and the realization of a binary phase-modulating spatial sound modulator that is based on an actively reconfigurable acoustic metasurface. We demonstrate useful functionalities including the creation of quiet zones and hotspots in a typical reverberating environment. Copyright © 2018 the Author(s). Published by PNAS.
Entropy in sound and vibration: towards a new paradigm.
Le Bot, A
2017-01-01
This paper describes a discussion on the method and the status of a statistical theory of sound and vibration, called statistical energy analysis (SEA). SEA is a simple theory of sound and vibration in elastic structures that applies when the vibrational energy is diffusely distributed. We show that SEA is a thermodynamical theory of sound and vibration, based on a law of exchange of energy analogous to the Clausius principle. We further investigate the notion of entropy in this context and discuss its meaning. We show that entropy is a measure of information lost in the passage from the classical theory of sound and vibration and SEA, its thermodynamical counterpart.
NASA Astrophysics Data System (ADS)
West, Eva; Wallin, Anita
2013-04-01
Learning abstract concepts such as sound often involves an ontological shift because to conceptualize sound transmission as a process of motion demands abandoning sound transmission as a transfer of matter. Thus, for students to be able to grasp and use a generalized model of sound transmission poses great challenges for them. This study involved 199 students aged 10-14. Their views about sound transmission were investigated before and after teaching by comparing their written answers about sound transfer in different media. The teaching was built on a research-based teaching-learning sequence (TLS), which was developed within a framework of design research. The analysis involved interpreting students' underlying theories of sound transmission, including the different conceptual categories that were found in their answers. The results indicated a shift in students' understandings from the use of a theory of matter before the intervention to embracing a theory of process afterwards. The described pattern was found in all groups of students irrespective of age. Thus, teaching about sound and sound transmission is fruitful already at the ages of 10-11. However, the older the students, the more advanced is their understanding of the process of motion. In conclusion, the use of a TLS about sound, hearing and auditory health promotes students' conceptualization of sound transmission as a process in all grades. The results also imply some crucial points in teaching and learning about the scientific content of sound.
Pitch features of environmental sounds
NASA Astrophysics Data System (ADS)
Yang, Ming; Kang, Jian
2016-07-01
A number of soundscape studies have suggested the need for suitable parameters for soundscape measurement, in addition to the conventional acoustic parameters. This paper explores the applicability of pitch features that are often used in music analysis and their algorithms to environmental sounds. Based on the existing alternative pitch algorithms for simulating the perception of the auditory system and simplified algorithms for practical applications in the areas of music and speech, the applicable algorithms have been determined, considering common types of sound in everyday soundscapes. Considering a number of pitch parameters, including pitch value, pitch strength, and percentage of audible pitches over time, different pitch characteristics of various environmental sounds have been shown. Among the four sound categories, i.e. water, wind, birdsongs, and urban sounds, generally speaking, both water and wind sounds have low pitch values and pitch strengths; birdsongs have high pitch values and pitch strengths; and urban sounds have low pitch values and a relatively wide range of pitch strengths.
A Method for the Estimation of p-Mode Parameters from Averaged Solar Oscillation Power Spectra
NASA Astrophysics Data System (ADS)
Reiter, J.; Rhodes, E. J., Jr.; Kosovichev, A. G.; Schou, J.; Scherrer, P. H.; Larson, T. P.
2015-04-01
A new fitting methodology is presented that is equally well suited for the estimation of low-, medium-, and high-degree mode parameters from m-averaged solar oscillation power spectra of widely differing spectral resolution. This method, which we call the “Windowed, MuLTiple-Peak, averaged-spectrum” or WMLTP Method, constructs a theoretical profile by convolving the weighted sum of the profiles of the modes appearing in the fitting box with the power spectrum of the window function of the observing run, using weights from a leakage matrix that takes into account observational and physical effects, such as the distortion of modes by solar latitudinal differential rotation. We demonstrate that the WMLTP Method makes substantial improvements in the inferences of the properties of the solar oscillations in comparison with a previous method, which employed a single profile to represent each spectral peak. We also present an inversion for the internal solar structure, which is based upon 6366 modes that we computed using the WMLTP method on the 66 day 2010 Solar and Heliospheric Observatory/MDI Dynamics Run. To improve both the numerical stability and reliability of the inversion, we developed a new procedure for the identification and correction of outliers in a frequency dataset. We present evidence for a pronounced departure of the sound speed in the outer half of the solar convection zone and in the subsurface shear layer from the radial sound speed profile contained in Model S of Christensen-Dalsgaard and his collaborators that existed in the rising phase of Solar Cycle 24 during mid-2010.
Yagci, Ahmet; Korkmaz, Yasemin Nur; Yagci, Filiz; Atilla, Aykan Onur; Buyuk, Suleyman Kutalmiş
2016-12-01
The aim of this study was to assess the effects of 3 luting agents (glass ionomer cement, compomer, and polycarboxylate cement) on white spot lesion formation in patients with full-coverage bonded acrylic splint expanders. White spot lesion formation was assessed with quantitative light-induced fluorescence. Full-coverage rapid maxillary expanders were cemented with glass ionomer cement, compomer, and polycarboxylate cement in groups 1, 2, and 3, respectively. A control group comprised patients who never had orthodontic treatment. Quantitative light-induced fluorescence images taken before and after rapid maxillary expansion treatment were analyzed for these parameters: the percentages of fluorescence loss with respect to the fluorescence of sound tooth tissue (ΔF) and maximum loss of fluorescence intensity in the whole lesion; lesion area with ΔF equal to less than a -5% threshold; and the percentage of fluorescence loss with respect to the fluorescence of sound tissue times the area that indicated lesion volume. All 3 groups showed statistically significantly greater demineralization than the control group. The 3 experimental groups differed from each other in half of the parameters calculated. Teeth in the polycarboxylate group developed the most white spot lesions. With the highest rate of white spot lesion formation, polycarboxylate cements should not be used for full-coverage bonded acrylic splint expanders. Compomers may be preferred over glass ionomer cements, based on the findings of this study. Copyright © 2016 American Association of Orthodontists. Published by Elsevier Inc. All rights reserved.
An intelligent artificial throat with sound-sensing ability based on laser induced graphene
Tao, Lu-Qi; Tian, He; Liu, Ying; Ju, Zhen-Yi; Pang, Yu; Chen, Yuan-Quan; Wang, Dan-Yang; Tian, Xiang-Guang; Yan, Jun-Chao; Deng, Ning-Qin; Yang, Yi; Ren, Tian-Ling
2017-01-01
Traditional sound sources and sound detectors are usually independent and discrete in the human hearing range. To minimize the device size and integrate it with wearable electronics, there is an urgent requirement of realizing the functional integration of generating and detecting sound in a single device. Here we show an intelligent laser-induced graphene artificial throat, which can not only generate sound but also detect sound in a single device. More importantly, the intelligent artificial throat will significantly assist for the disabled, because the simple throat vibrations such as hum, cough and scream with different intensity or frequency from a mute person can be detected and converted into controllable sounds. Furthermore, the laser-induced graphene artificial throat has the advantage of one-step fabrication, high efficiency, excellent flexibility and low cost, and it will open practical applications in voice control, wearable electronics and many other areas. PMID:28232739
An intelligent artificial throat with sound-sensing ability based on laser induced graphene.
Tao, Lu-Qi; Tian, He; Liu, Ying; Ju, Zhen-Yi; Pang, Yu; Chen, Yuan-Quan; Wang, Dan-Yang; Tian, Xiang-Guang; Yan, Jun-Chao; Deng, Ning-Qin; Yang, Yi; Ren, Tian-Ling
2017-02-24
Traditional sound sources and sound detectors are usually independent and discrete in the human hearing range. To minimize the device size and integrate it with wearable electronics, there is an urgent requirement of realizing the functional integration of generating and detecting sound in a single device. Here we show an intelligent laser-induced graphene artificial throat, which can not only generate sound but also detect sound in a single device. More importantly, the intelligent artificial throat will significantly assist for the disabled, because the simple throat vibrations such as hum, cough and scream with different intensity or frequency from a mute person can be detected and converted into controllable sounds. Furthermore, the laser-induced graphene artificial throat has the advantage of one-step fabrication, high efficiency, excellent flexibility and low cost, and it will open practical applications in voice control, wearable electronics and many other areas.
NASA Astrophysics Data System (ADS)
Yahya, I.; Kusuma, J. I.; Harjana; Kristiani, R.; Hanina, R.
2016-02-01
This paper emphasizes the influence of tubular shaped microresonators phononic crystal insertion on the sound absorption coefficient of profiled sound absorber. A simple cubic and two different bodies centered cubic phononic crystal lattice model were analyzed in a laboratory test procedure. The experiment was conducted by using transfer function based two microphone impedance tube method refer to ASTM E-1050-98. The results show that sound absorption coefficient increase significantly at the mid and high-frequency band (600 - 700 Hz) and (1 - 1.6 kHz) when tubular shaped microresonator phononic crystal inserted into the tested sound absorber element. The increment phenomena related to multi-resonance effect that occurs when sound waves propagate through the phononic crystal lattice model that produce multiple reflections and scattering in mid and high-frequency band which increases the sound absorption coefficient accordingly
An intelligent artificial throat with sound-sensing ability based on laser induced graphene
NASA Astrophysics Data System (ADS)
Tao, Lu-Qi; Tian, He; Liu, Ying; Ju, Zhen-Yi; Pang, Yu; Chen, Yuan-Quan; Wang, Dan-Yang; Tian, Xiang-Guang; Yan, Jun-Chao; Deng, Ning-Qin; Yang, Yi; Ren, Tian-Ling
2017-02-01
Traditional sound sources and sound detectors are usually independent and discrete in the human hearing range. To minimize the device size and integrate it with wearable electronics, there is an urgent requirement of realizing the functional integration of generating and detecting sound in a single device. Here we show an intelligent laser-induced graphene artificial throat, which can not only generate sound but also detect sound in a single device. More importantly, the intelligent artificial throat will significantly assist for the disabled, because the simple throat vibrations such as hum, cough and scream with different intensity or frequency from a mute person can be detected and converted into controllable sounds. Furthermore, the laser-induced graphene artificial throat has the advantage of one-step fabrication, high efficiency, excellent flexibility and low cost, and it will open practical applications in voice control, wearable electronics and many other areas.
PROTAX-Sound: A probabilistic framework for automated animal sound identification
Somervuo, Panu; Ovaskainen, Otso
2017-01-01
Autonomous audio recording is stimulating new field in bioacoustics, with a great promise for conducting cost-effective species surveys. One major current challenge is the lack of reliable classifiers capable of multi-species identification. We present PROTAX-Sound, a statistical framework to perform probabilistic classification of animal sounds. PROTAX-Sound is based on a multinomial regression model, and it can utilize as predictors any kind of sound features or classifications produced by other existing algorithms. PROTAX-Sound combines audio and image processing techniques to scan environmental audio files. It identifies regions of interest (a segment of the audio file that contains a vocalization to be classified), extracts acoustic features from them and compares with samples in a reference database. The output of PROTAX-Sound is the probabilistic classification of each vocalization, including the possibility that it represents species not present in the reference database. We demonstrate the performance of PROTAX-Sound by classifying audio from a species-rich case study of tropical birds. The best performing classifier achieved 68% classification accuracy for 200 bird species. PROTAX-Sound improves the classification power of current techniques by combining information from multiple classifiers in a manner that yields calibrated classification probabilities. PMID:28863178
PROTAX-Sound: A probabilistic framework for automated animal sound identification.
de Camargo, Ulisses Moliterno; Somervuo, Panu; Ovaskainen, Otso
2017-01-01
Autonomous audio recording is stimulating new field in bioacoustics, with a great promise for conducting cost-effective species surveys. One major current challenge is the lack of reliable classifiers capable of multi-species identification. We present PROTAX-Sound, a statistical framework to perform probabilistic classification of animal sounds. PROTAX-Sound is based on a multinomial regression model, and it can utilize as predictors any kind of sound features or classifications produced by other existing algorithms. PROTAX-Sound combines audio and image processing techniques to scan environmental audio files. It identifies regions of interest (a segment of the audio file that contains a vocalization to be classified), extracts acoustic features from them and compares with samples in a reference database. The output of PROTAX-Sound is the probabilistic classification of each vocalization, including the possibility that it represents species not present in the reference database. We demonstrate the performance of PROTAX-Sound by classifying audio from a species-rich case study of tropical birds. The best performing classifier achieved 68% classification accuracy for 200 bird species. PROTAX-Sound improves the classification power of current techniques by combining information from multiple classifiers in a manner that yields calibrated classification probabilities.
Aeroacoustic analysis of the human phonation process based on a hybrid acoustic PIV approach
NASA Astrophysics Data System (ADS)
Lodermeyer, Alexander; Tautz, Matthias; Becker, Stefan; Döllinger, Michael; Birk, Veronika; Kniesburges, Stefan
2018-01-01
The detailed analysis of sound generation in human phonation is severely limited as the accessibility to the laryngeal flow region is highly restricted. Consequently, the physical basis of the underlying fluid-structure-acoustic interaction that describes the primary mechanism of sound production is not yet fully understood. Therefore, we propose the implementation of a hybrid acoustic PIV procedure to evaluate aeroacoustic sound generation during voice production within a synthetic larynx model. Focusing on the flow field downstream of synthetic, aerodynamically driven vocal folds, we calculated acoustic source terms based on the velocity fields obtained by time-resolved high-speed PIV applied to the mid-coronal plane. The radiation of these sources into the acoustic far field was numerically simulated and the resulting acoustic pressure was finally compared with experimental microphone measurements. We identified the tonal sound to be generated downstream in a small region close to the vocal folds. The simulation of the sound propagation underestimated the tonal components, whereas the broadband sound was well reproduced. Our results demonstrate the feasibility to locate aeroacoustic sound sources inside a synthetic larynx using a hybrid acoustic PIV approach. Although the technique employs a 2D-limited flow field, it accurately reproduces the basic characteristics of the aeroacoustic field in our larynx model. In future studies, not only the aeroacoustic mechanisms of normal phonation will be assessable, but also the sound generation of voice disorders can be investigated more profoundly.
Geoelectrical characterization by joint inversion of VES/TEM in Paraná basin, Brazil
NASA Astrophysics Data System (ADS)
Bortolozo, C. A.; Couto, M. A.; Almeida, E. R.; Porsani, J. L.; Santos, F. M.
2012-12-01
For many years electrical (DC) and transient electromagnetic (TEM) soundings have been used in a great number of environmental, hydrological and mining exploration studies. The data of both methods are interpreted usually by individual 1D models resulting in many cases in ambiguous models. This can be explained by how the two different methodologies sample the subsurface. The vertical electrical sounding (VES) is good on marking very resistive structures, while the transient electromagnetic sounding (TEM) is very sensitive to map conductive structures. Another characteristic is that VES is more sensitive to shallow structures, while TEM soundings can reach deeper structures. A Matlab program for joint inversion of VES and TEM soundings, by using CRS algorithm was developed aiming explore the best of the both methods. Initially, the algorithm was tested with synthetic data and after it was used to invert experimental data from Paraná sedimentary basin. We present the results of a re-interpretation of 46 VES/TEM soundings data set acquired in Bebedouro region in São Paulo State - Brazil. The previous interpretation was based in geoelectrical models obtained by single inversion of the VES and TEM soundings. In this work we present the results with single inversion of VES and TEM sounding inverted by the Curupira Program and a new interpretation based in the joint inversion of both methodologies. The goal is increase the accuracy in determining the underground structures. As a result a new geoelectrical model of the region is obtained.
[Perception and selectivity of sound duration in the central auditory midbrain].
Wang, Xin; Li, An-An; Wu, Fei-Jian
2010-08-25
Sound duration plays important role in acoustic communication. Information of acoustic signal is mainly encoded in the amplitude and frequency spectrum of different durations. Duration selective neurons exist in the central auditory system including inferior colliculus (IC) of frog, bat, mouse and chinchilla, etc., and they are important in signal recognition and feature detection. Two generally accepted models, which are "coincidence detector model" and "anti-coincidence detector model", have been raised to explain the mechanism of neural selective responses to sound durations based on the study of IC neurons in bats. Although they are different in details, they both emphasize the importance of synaptic integration of excitatory and inhibitory inputs, and are able to explain the responses of most duration-selective neurons. However, both of the hypotheses need to be improved since other sound parameters, such as spectral pattern, amplitude and repetition rate, could affect the duration selectivity of the neurons. The dynamic changes of sound parameters are believed to enable the animal to effectively perform recognition of behavior related acoustic signals. Under free field sound stimulation, we analyzed the neural responses in the IC and auditory cortex of mouse and bat to sounds with different duration, frequency and amplitude, using intracellular or extracellular recording techniques. Based on our work and previous studies, this article reviews the properties of duration selectivity in central auditory system and discusses the mechanisms of duration selectivity and the effect of other sound parameters on the duration coding of auditory neurons.
NASA Astrophysics Data System (ADS)
Wang, Y. S.; Shen, G. Q.; Xing, Y. F.
2014-03-01
Based on the artificial neural network (ANN) technique, an objective sound quality evaluation (SQE) model for synthesis annoyance of vehicle interior noises is presented in this paper. According to the standard named GB/T18697, firstly, the interior noises under different working conditions of a sample vehicle are measured and saved in a noise database. Some mathematical models for loudness, sharpness and roughness of the measured vehicle noises are established and performed by Matlab programming. Sound qualities of the vehicle interior noises are also estimated by jury tests following the anchored semantic differential (ASD) procedure. Using the objective and subjective evaluation results, furthermore, an ANN-based model for synthetical annoyance evaluation of vehicle noises, so-called ANN-SAE, is developed. Finally, the ANN-SAE model is proved by some verification tests with the leave-one-out algorithm. The results suggest that the proposed ANN-SAE model is accurate and effective and can be directly used to estimate sound quality of the vehicle interior noises, which is very helpful for vehicle acoustical designs and improvements. The ANN-SAE approach may be extended to deal with other sound-related fields for product quality evaluations in SQE engineering.
Cell type-specific suppression of mechanosensitive genes by audible sound stimulation.
Kumeta, Masahiro; Takahashi, Daiji; Takeyasu, Kunio; Yoshimura, Shige H
2018-01-01
Audible sound is a ubiquitous environmental factor in nature that transmits oscillatory compressional pressure through the substances. To investigate the property of the sound as a mechanical stimulus for cells, an experimental system was set up using 94.0 dB sound which transmits approximately 10 mPa pressure to the cultured cells. Based on research on mechanotransduction and ultrasound effects on cells, gene responses to the audible sound stimulation were analyzed by varying several sound parameters: frequency, wave form, composition, and exposure time. Real-time quantitative PCR analyses revealed a distinct suppressive effect for several mechanosensitive and ultrasound-sensitive genes that were triggered by sounds. The effect was clearly observed in a wave form- and pressure level-specific manner, rather than the frequency, and persisted for several hours. At least two mechanisms are likely to be involved in this sound response: transcriptional control and RNA degradation. ST2 stromal cells and C2C12 myoblasts exhibited a robust response, whereas NIH3T3 cells were partially and NB2a neuroblastoma cells were completely insensitive, suggesting a cell type-specific response to sound. These findings reveal a cell-level systematic response to audible sound and uncover novel relationships between life and sound.
Cell type-specific suppression of mechanosensitive genes by audible sound stimulation
Takahashi, Daiji; Takeyasu, Kunio; Yoshimura, Shige H.
2018-01-01
Audible sound is a ubiquitous environmental factor in nature that transmits oscillatory compressional pressure through the substances. To investigate the property of the sound as a mechanical stimulus for cells, an experimental system was set up using 94.0 dB sound which transmits approximately 10 mPa pressure to the cultured cells. Based on research on mechanotransduction and ultrasound effects on cells, gene responses to the audible sound stimulation were analyzed by varying several sound parameters: frequency, wave form, composition, and exposure time. Real-time quantitative PCR analyses revealed a distinct suppressive effect for several mechanosensitive and ultrasound-sensitive genes that were triggered by sounds. The effect was clearly observed in a wave form- and pressure level-specific manner, rather than the frequency, and persisted for several hours. At least two mechanisms are likely to be involved in this sound response: transcriptional control and RNA degradation. ST2 stromal cells and C2C12 myoblasts exhibited a robust response, whereas NIH3T3 cells were partially and NB2a neuroblastoma cells were completely insensitive, suggesting a cell type-specific response to sound. These findings reveal a cell-level systematic response to audible sound and uncover novel relationships between life and sound. PMID:29385174
SoundCompass: A Distributed MEMS Microphone Array-Based Sensor for Sound Source Localization
Tiete, Jelmer; Domínguez, Federico; da Silva, Bruno; Segers, Laurent; Steenhaut, Kris; Touhafi, Abdellah
2014-01-01
Sound source localization is a well-researched subject with applications ranging from localizing sniper fire in urban battlefields to cataloging wildlife in rural areas. One critical application is the localization of noise pollution sources in urban environments, due to an increasing body of evidence linking noise pollution to adverse effects on human health. Current noise mapping techniques often fail to accurately identify noise pollution sources, because they rely on the interpolation of a limited number of scattered sound sensors. Aiming to produce accurate noise pollution maps, we developed the SoundCompass, a low-cost sound sensor capable of measuring local noise levels and sound field directionality. Our first prototype is composed of a sensor array of 52 Microelectromechanical systems (MEMS) microphones, an inertial measuring unit and a low-power field-programmable gate array (FPGA). This article presents the SoundCompass’s hardware and firmware design together with a data fusion technique that exploits the sensing capabilities of the SoundCompass in a wireless sensor network to localize noise pollution sources. Live tests produced a sound source localization accuracy of a few centimeters in a 25-m2 anechoic chamber, while simulation results accurately located up to five broadband sound sources in a 10,000-m2 open field. PMID:24463431
Sight over sound in the judgment of music performance.
Tsay, Chia-Jung
2013-09-03
Social judgments are made on the basis of both visual and auditory information, with consequential implications for our decisions. To examine the impact of visual information on expert judgment and its predictive validity for performance outcomes, this set of seven experiments in the domain of music offers a conservative test of the relative influence of vision versus audition. People consistently report that sound is the most important source of information in evaluating performance in music. However, the findings demonstrate that people actually depend primarily on visual information when making judgments about music performance. People reliably select the actual winners of live music competitions based on silent video recordings, but neither musical novices nor professional musicians were able to identify the winners based on sound recordings or recordings with both video and sound. The results highlight our natural, automatic, and nonconscious dependence on visual cues. The dominance of visual information emerges to the degree that it is overweighted relative to auditory information, even when sound is consciously valued as the core domain content.
Development of a hybrid wave based-transfer matrix model for sound transmission analysis.
Dijckmans, A; Vermeir, G
2013-04-01
In this paper, a hybrid wave based-transfer matrix model is presented that allows for the investigation of the sound transmission through finite multilayered structures placed between two reverberant rooms. The multilayered structure may consist of an arbitrary configuration of fluid, elastic, or poro-elastic layers. The field variables (structural displacements and sound pressures) are expanded in terms of structural and acoustic wave functions. The boundary and continuity conditions in the rooms determine the participation factors in the pressure expansions. The displacement of the multilayered structure is determined by the mechanical impedance matrix, which gives a relation between the pressures and transverse displacements at both sides of the structure. The elements of this matrix are calculated with the transfer matrix method. First, the hybrid model is numerically validated. Next a comparison is made with sound transmission loss measurements of a hollow brick wall and a sandwich panel. Finally, numerical simulations show the influence of structural damping, room dimensions and plate dimensions on the sound transmission loss of multilayered structures.
Orban, David A; Soltis, Joseph; Perkins, Lori; Mellen, Jill D
2017-05-01
A clear need for evidence-based animal management in zoos and aquariums has been expressed by industry leaders. Here, we show how individual animal welfare monitoring can be combined with measurement of environmental conditions to inform science-based animal management decisions. Over the last several years, Disney's Animal Kingdom® has been undergoing significant construction and exhibit renovation, warranting institution-wide animal welfare monitoring. Animal care and science staff developed a model that tracked animal keepers' daily assessments of an animal's physical health, behavior, and responses to husbandry activity; these data were matched to different external stimuli and environmental conditions, including sound levels. A case study of a female giant anteater and her environment is presented to illustrate how this process worked. Associated with this case, several sound-reducing barriers were tested for efficacy in mitigating sound. Integrating daily animal welfare assessment with environmental monitoring can lead to a better understanding of animals and their sensory environment and positively impact animal welfare. © 2017 Wiley Periodicals, Inc.
Sight over sound in the judgment of music performance
Tsay, Chia-Jung
2013-01-01
Social judgments are made on the basis of both visual and auditory information, with consequential implications for our decisions. To examine the impact of visual information on expert judgment and its predictive validity for performance outcomes, this set of seven experiments in the domain of music offers a conservative test of the relative influence of vision versus audition. People consistently report that sound is the most important source of information in evaluating performance in music. However, the findings demonstrate that people actually depend primarily on visual information when making judgments about music performance. People reliably select the actual winners of live music competitions based on silent video recordings, but neither musical novices nor professional musicians were able to identify the winners based on sound recordings or recordings with both video and sound. The results highlight our natural, automatic, and nonconscious dependence on visual cues. The dominance of visual information emerges to the degree that it is overweighted relative to auditory information, even when sound is consciously valued as the core domain content. PMID:23959902
Zhao, Sipei; Qiu, Xiaojun; Cheng, Jianchun
2015-09-01
This paper proposes a different method for calculating a sound field diffracted by a rigid barrier based on the integral equation method, where a virtual boundary is assumed above the rigid barrier to divide the whole space into two subspaces. Based on the Kirchhoff-Helmholtz equation, the sound field in each subspace is determined with the source inside and the boundary conditions on the surface, and then the diffracted sound field is obtained by using the continuation conditions on the virtual boundary. Simulations are carried out to verify the feasibility of the proposed method. Compared to the MacDonald method and other existing methods, the proposed method is a rigorous solution for whole space and is also much easier to understand.
How do "mute" cicadas produce their calling songs?
Luo, Changqing; Wei, Cong; Nansen, Christian
2015-01-01
Insects have evolved a variety of structures and mechanisms to produce sounds, which are used for communication both within and between species. Among acoustic insects, cicada males are particularly known for their loud and diverse sounds which function importantly in communication. The main method of sound production in cicadas is the tymbal mechanism, and a relative small number of cicada species possess both tymbal and stridulatory organs. However, cicadas of the genus Karenia do not have any specialized sound-producing structures, so they are referred to as "mute". This denomination is quite misleading, as they indeed produce sounds. Here, we investigate the sound-producing mechanism and acoustic communication of the "mute" cicada, Karenia caelatata, and discover a new sound-production mechanism for cicadas: i.e., K. caelatata produces impact sounds by banging the forewing costa against the operculum. The temporal, frequency and amplitude characteristics of the impact sounds are described. Morphological studies and reflectance-based analyses reveal that the structures involved in sound production of K. caelatata (i.e., forewing, operculum, cruciform elevation, and wing-holding groove on scutellum) are all morphologically modified. Acoustic playback experiments and behavioral observations suggest that the impact sounds of K. caelatata are used in intraspecific communication and function as calling songs. The new sound-production mechanism expands our knowledge on the diversity of acoustic signaling behavior in cicadas and further underscores the need for more bioacoustic studies on cicadas which lack tymbal mechanism.
Statistical Analysis for Subjective and Objective Evaluations of Dental Drill Sounds
Yamada, Tomomi; Kuwano, Sonoko; Ebisu, Shigeyuki; Hayashi, Mikako
2016-01-01
The sound produced by a dental air turbine handpiece (dental drill) can markedly influence the sound environment in a dental clinic. Indeed, many patients report that the sound of a dental drill elicits an unpleasant feeling. Although several manufacturers have attempted to reduce the sound pressure levels produced by dental drills during idling based on ISO 14457, the sound emitted by such drills under active drilling conditions may negatively influence the dental clinic sound environment. The physical metrics related to the unpleasant impressions associated with dental drill sounds have not been determined. In the present study, psychological measurements of dental drill sounds were conducted with the aim of facilitating improvement of the sound environment at dental clinics. Specifically, we examined the impressions elicited by the sounds of 12 types of dental drills in idling and drilling conditions using a semantic differential. The analysis revealed that the impressions of dental drill sounds varied considerably between idling and drilling conditions and among the examined drills. This finding suggests that measuring the sound of a dental drill in idling conditions alone may be insufficient for evaluating the effects of the sound. We related the results of the psychological evaluations to those of measurements of the physical metrics of equivalent continuous A-weighted sound pressure levels (LAeq) and sharpness. Factor analysis indicated that impressions of the dental drill sounds consisted of two factors: “metallic and unpleasant” and “powerful”. LAeq had a strong relationship with “powerful impression”, calculated sharpness was positively related to “metallic impression”, and “unpleasant impression” was predicted by the combination of both LAeq and calculated sharpness. The present analyses indicate that, in addition to a reduction in sound pressure level, refining the frequency components of dental drill sounds is important for creating a comfortable sound environment in dental clinics. PMID:27462903
Vocal Imitations of Non-Vocal Sounds
Houix, Olivier; Voisin, Frédéric; Misdariis, Nicolas; Susini, Patrick
2016-01-01
Imitative behaviors are widespread in humans, in particular whenever two persons communicate and interact. Several tokens of spoken languages (onomatopoeias, ideophones, and phonesthemes) also display different degrees of iconicity between the sound of a word and what it refers to. Thus, it probably comes at no surprise that human speakers use a lot of imitative vocalizations and gestures when they communicate about sounds, as sounds are notably difficult to describe. What is more surprising is that vocal imitations of non-vocal everyday sounds (e.g. the sound of a car passing by) are in practice very effective: listeners identify sounds better with vocal imitations than with verbal descriptions, despite the fact that vocal imitations are inaccurate reproductions of a sound created by a particular mechanical system (e.g. a car driving by) through a different system (the voice apparatus). The present study investigated the semantic representations evoked by vocal imitations of sounds by experimentally quantifying how well listeners could match sounds to category labels. The experiment used three different types of sounds: recordings of easily identifiable sounds (sounds of human actions and manufactured products), human vocal imitations, and computational “auditory sketches” (created by algorithmic computations). The results show that performance with the best vocal imitations was similar to the best auditory sketches for most categories of sounds, and even to the referent sounds themselves in some cases. More detailed analyses showed that the acoustic distance between a vocal imitation and a referent sound is not sufficient to account for such performance. Analyses suggested that instead of trying to reproduce the referent sound as accurately as vocally possible, vocal imitations focus on a few important features, which depend on each particular sound category. These results offer perspectives for understanding how human listeners store and access long-term sound representations, and sets the stage for the development of human-computer interfaces based on vocalizations. PMID:27992480
Noise Radiation Of A Strongly Pulsating Tailpipe Exhaust
NASA Astrophysics Data System (ADS)
Peizi, Li; Genhua, Dai; Zhichi, Zhu
1993-11-01
The method of characteristics is used to solve the problem of the propagation of a strongly pulsating flow in an exhaust system tailpipe. For a strongly pulsating exhaust, the flow may shock at the pipe's open end at some point in a pulsating where the flow pressure exceeds its critical value. The method fails if one insists on setting the flow pressure equal to the atmospheric pressure as the pipe end boundary condition. To solve the problem, we set the Mach number equal to 1 as the boundary condition when the flow pressure exceeds its critical value. For a strongly pulsating flow, the fluctuations of flow variables may be much higher than their respective time averages. Therefore, the acoustic radiation method would fail in the computation of the noise radiation from the pipe's open end. We simulate the exhaust flow out of the open end as a simple sound source to compute the noise radiation, which has been successfully applied in reference [1]. The simple sound source strength is proportional to the volume acceleration of exhaust gas. Also computed is the noise radiation from the turbulence of the exhaust flow, as was done in reference [1]. Noise from a reciprocating valve simulator has been treated in detail. The radiation efficiency is very low for the pressure range considered and is about 10 -5. The radiation efficiency coefficient increases with the square of the frequency. Computation of the pipe length dependence of the noise radiation and mass flux allows us to design a suitable length for an aerodynamic noise generator or a reciprocating internal combustion engine. For the former, powerful noise radiation is preferable. For the latter, maximum mass flux is desired because a freer exhaust is preferable.
Decision strategies of hearing-impaired listeners in spectral shape discrimination
NASA Astrophysics Data System (ADS)
Lentz, Jennifer J.; Leek, Marjorie R.
2002-03-01
The ability to discriminate between sounds with different spectral shapes was evaluated for normal-hearing and hearing-impaired listeners. Listeners detected a 920-Hz tone added in phase to a single component of a standard consisting of the sum of five tones spaced equally on a logarithmic frequency scale ranging from 200 to 4200 Hz. An overall level randomization of 10 dB was either present or absent. In one subset of conditions, the no-perturbation conditions, the standard stimulus was the sum of equal-amplitude tones. In the perturbation conditions, the amplitudes of the components within a stimulus were randomly altered on every presentation. For both perturbation and no-perturbation conditions, thresholds for the detection of the 920-Hz tone were measured to compare sensitivity to changes in spectral shape between normal-hearing and hearing-impaired listeners. To assess whether hearing-impaired listeners relied on different regions of the spectrum to discriminate between sounds, spectral weights were estimated from the perturbed standards by correlating the listener's responses with the level differences per component across two intervals of a two-alternative forced-choice task. Results showed that hearing-impaired and normal-hearing listeners had similar sensitivity to changes in spectral shape. On average, across-frequency correlation functions also were similar for both groups of listeners, suggesting that as long as all components are audible and well separated in frequency, hearing-impaired listeners can use information across frequency as well as normal-hearing listeners. Analysis of the individual data revealed, however, that normal-hearing listeners may be better able to adopt optimal weighting schemes. This conclusion is only tentative, as differences in internal noise may need to be considered to interpret the results obtained from weighting studies between normal-hearing and hearing-impaired listeners.
Objective and perceptual comparisons of two bluetooth hearing aid assistive devices.
Clark, Jackie L; Pustejovsky, Carmen; Vanneste, Sven
2017-08-01
With the advent of Bluetooth technology, many of the assistive listening devices for hearing have become manufacturer specific, with little objective information about the performance provided. Thirty native English-speaking adults (mean age 29.8) with normal hearing were tested pseudo-randomly with two major hearing aid manufacturers' proprietary Bluetooth connectivity devices paired to the accompanying manufacturer's specific hearing aids. Sentence recognition performance was objectively measured for each system with signals transmitted via a land-line to the same iPhone in two conditions. There was a significant effect of participant's performance according to listening condition. There was no significant effect between device manufacturers according to listening condition, but there was a significant effect in participant's perception of "quality of sound". Despite differences in signal transmission for each devise, when worn by participants both the systems performed equally. In fact, participants expressed personal preferences for specific technology that was largely due to their perceived quality of sound while listening to recorded signals. While further research is necessary to investigate other measures of benefit for Bluetooth connectivity devices, preliminary data suggest that in order to ensure comfort and compatibility, not only should objective measures of the patient benefit be completed, but also assessing the patient's perception of benefit is equally important. Implications for Rehabilitation All professionals who work with individuals with hearing loss, become aware of the differences in the multiple choices for assistive technology readily available for hearing loss. With the ever growing dispensing of Bluetooth connectivity devices coupled to hearing aids, there is an increased burden to determine whether performance differences could exist between manufacturers. There is a growing need to investigate other measures of benefit for Bluetooth hearing aid connectivity devices that not only include objective measures, but also patient perception of benefit.
Effects of sound intensity on temporal properties of inhibition in the pallid bat auditory cortex.
Razak, Khaleel A
2013-01-01
Auditory neurons in bats that use frequency modulated (FM) sweeps for echolocation are selective for the behaviorally-relevant rates and direction of frequency change. Such selectivity arises through spectrotemporal interactions between excitatory and inhibitory components of the receptive field. In the pallid bat auditory system, the relationship between FM sweep direction/rate selectivity and spectral and temporal properties of sideband inhibition have been characterized. Of note is the temporal asymmetry in sideband inhibition, with low-frequency inhibition (LFI) exhibiting faster arrival times compared to high-frequency inhibition (HFI). Using the two-tone inhibition over time (TTI) stimulus paradigm, this study investigated the interactions between two sound parameters in shaping sideband inhibition: intensity and time. Specifically, the impact of changing relative intensities of the excitatory and inhibitory tones on arrival time of inhibition was studied. Using this stimulation paradigm, single unit data from the auditory cortex of pentobarbital-anesthetized cortex show that the threshold for LFI is on average ~8 dB lower than HFI. For equal intensity tones near threshold, LFI is stronger than HFI. When the inhibitory tone intensity is increased further from threshold, the strength asymmetry decreased. The temporal asymmetry in LFI vs. HFI arrival time is strongest when the excitatory and inhibitory tones are of equal intensities or if excitatory tone is louder. As inhibitory tone intensity is increased, temporal asymmetry decreased suggesting that the relative magnitude of excitatory and inhibitory inputs shape arrival time of inhibition and FM sweep rate and direction selectivity. Given that most FM bats use downward sweeps as echolocation calls, a similar asymmetry in threshold and strength of LFI vs. HFI may be a general adaptation to enhance direction selectivity while maintaining sweep-rate selective responses to downward sweeps.
The transmission of finite amplitude sound beam in multi-layered biological media
NASA Astrophysics Data System (ADS)
Liu, Xiaozhou; Li, Junlun; Yin, Chang; Gong, Xiufen; Zhang, Dong; Xue, Honghui
2007-02-01
Based on the Khokhlov Zabolotskaya Kuznetsov (KZK) equation, a model in the frequency domain is given to describe the transmission of finite amplitude sound beam in multi-layered biological media. Favorable agreement between the theoretical analyses and the measured results shows this approach could effectively describe the transmission of finite amplitude sound wave in multi-layered biological media.
ERIC Educational Resources Information Center
Zeng, Liang; Smith, Chris; Poelzer, G. Herold; Rodriguez, Jennifer; Corpuz, Edgar; Yanev, George
2014-01-01
In our pilot studies, we found that many introductory physics textbook illustrations with supporting text for sound standing waves of air columns in open-open, open-closed, and closed-closed pipes inhibit student understanding of sound standing wave phenomena due to student misunderstanding of how air molecules move within these pipes. Based on…
Multidimensional Approach to the Development of a Mandarin Chinese-Oriented Sound Test
ERIC Educational Resources Information Center
Hung, Yu-Chen; Lin, Chun-Yi; Tsai, Li-Chiun; Lee, Ya-Jung
2016-01-01
Purpose: Because the Ling six-sound test is based on American English phonemes, it can yield unreliable results when administered to non-English speakers. In this study, we aimed to improve specifically the diagnostic palette for Mandarin Chinese users by developing an adapted version of the Ling six-sound test. Method: To determine the set of…
Effect of sound on boundary layer stability
NASA Technical Reports Server (NTRS)
Saric, William S. (Principal Investigator); Spencer, Shelly Anne
1993-01-01
Experiments are conducted in the Arizona State University Unsteady Wind Tunnel with a zero-pressure-gradient flat-plate model that has a 67:1 elliptical leading edge. Boundary-layer measurements are made of the streamwise fluctuating-velocity component in order to identify the amplified T-S waves that are forced by downstream-travelling, sound waves. Measurements are taken with circular 3-D roughness elements placed at the Branch 1 neutral stability point for the frequency under consideration, and then with the roughness element downstream of Branch 1. These roughness elements have a principal chord dimension equal to 2(lambda)(sub TS)/pi, of the T-S waves under study and are 'stacked' in order to resemble a Gaussian height distribution. Measurements taken just downstream of the roughness (with leading-edge T-S waves, surface roughness T-S waves, instrumentation sting vibrations and the Stokes wave subtracted) show the generation of 3-D-T-S waves, but not in the characteristic heart-shaped disturbance field predicted by 3-D asymptotic theory. Maximum disturbance amplitudes are found on the roughness centerline. However, some near-field characteristics predicted by numerical modelling are observed.
Effect of sound on boundary layer stability
NASA Technical Reports Server (NTRS)
Saric, William S.; Spencer, Shelly Anne
1993-01-01
Experiments are conducted in the Arizona State University Unsteady Wind Tunnel with a zero-pressure-gradient flat-plate model that has a 67:1 elliptical leading edge. Boundary-layer measurements are made of the streamwise fluctuating-velocity component in order to identify the amplified T-S waves that are forced by downstream-traveling sound waves. Measurements are taken with circular 3-D roughness elements placed at the Branch 1 neutral stability point for the frequency under consideration, and then with the roughness element downstream of Branch 1. These roughness elements have a principal chord dimension equal to 2 lambda(sub TS)/pi of the T-S waves under study and are 'stacked' in order to resemble a Gaussian height distribution. Measurements taken just downstream of the roughness (with leading-edge T-S waves, surface roughness T-S waves, instrumentation sting vibrations, and the Stokes wave subtracted) show the generation of 3-D T-S waves, but not in the characteristic heart-shaped disturbance field predicted by 3-D asymptotic theory. Maximum disturbance amplitudes are found on the roughness centerline. However, some near-field characteristics predicted by numerical modeling are observed.
A passive means for cancellation of structurally radiated tones.
Zapfe, Jeffrey A; Ungar, Eric E
2003-01-01
The concept of cancellation of constant-frequency sound radiated from a vibrating surface by means of an attached mechanical oscillator is discussed. It is observed that the mass of a mechanical oscillator whose spring is attached to the vibrating surface will vibrate at comparatively large amplitudes and out of phase with that surface, provided that the surface vibrates at a frequency that is slightly higher than the oscillator's natural frequency. From this observation it is concluded that an oscillator's mass with a relatively small surface area can produce a volume velocity that is equal and opposite to that of the vibrating surface, resulting in cancellation of the sound radiated from the surface. Practical considerations in the design of such an oscillator are discussed, and the canceling performance from oscillators consisting of edge-supported circular disks is analyzed. An experimental canceling oscillator consisting of an edge-supported disk is described, and measurements made with this disk attached to a piston are shown to be in good agreement with analytical predictions. A tonal noise reduction exceeding 20 dB was demonstrated experimentally.
Hayes, Michael C.; Hays, Richard; Rubin, Stephen P.; Chase, Dorothy M.; Hallock, Molly; Cook-Tabor, Carrie; Luzier, Christina W.; Moser, Mary L.
2013-01-01
Lamprey populations are in decline worldwide and the status of Pacific lamprey (Entosphenus tridentatus) is a topic of current interest. They and other lamprey species cycle nutrients and serve as prey in riverine ecosystems. To determine the current distribution of Pacific lamprey in major watersheds flowing into Puget Sound, Washington, we sampled lamprey captured during salmonid smolt monitoring that occurred from late winter to mid-summer. We found Pacific lamprey in 12 of 18 watersheds and they were most common in southern Puget Sound watersheds and in watersheds draining western Puget Sound (Hood Canal). Two additional species, western brook lamprey (Lampetra richardsoni) and river lamprey (L. ayresii) were more common in eastern Puget Sound watersheds. Few Pacific lamprey macrophthalmia were found, suggesting that the majority of juveniles migrated seaward during other time periods. In addition, “dwarf” adult Pacific lamprey (< 300 mm) were observed in several watersheds and may represent an alternate life history for some Puget Sound populations. Based on genetic data, the use of visual techniques to identify lamprey ammocoetes as Entosphenus or Lampetra was successful for 97% (34 of 35) of the samples we evaluated.
He, Zhixue; Li, Xiang; Luo, Ming; Hu, Rong; Li, Cai; Qiu, Ying; Fu, Songnian; Yang, Qi; Yu, Shaohua
2016-05-02
We propose and experimentally demonstrate two independent component analysis (ICA) based channel equalizers (CEs) for 6 × 6 MIMO-OFDM transmission over few-mode fiber. Compared with the conventional channel equalizer based on training symbols (TSs-CE), the proposed two ICA-based channel equalizers (ICA-CE-I and ICA-CE-II) can achieve comparable performances, while requiring much less training symbols. Consequently, the overheads for channel equalization can be substantially reduced from 13.7% to 0.4% and 2.6%, respectively. Meanwhile, we also experimentally investigate the convergence speed of the proposed ICA-based CEs.
Software-Based Scoring and Sound Design: An Introductory Guide for Music Technology Instruction
ERIC Educational Resources Information Center
Walzer, Daniel A.
2016-01-01
This article explores the creative function of virtual instruments, sequencers, loops, and software-based synthesizers to introduce basic scoring and sound design concepts for visual media in an introductory music technology course. Using digital audio workstations with user-focused and configurable options, novice composers can hone a broad range…
ERIC Educational Resources Information Center
Stickney, Jeff Alan
2009-01-01
Comparing the early, analytic attempt to define "sound" teaching with the current use of criteria-based rating schemes, Jeff Stickney turns to Wittgenstein's holistic, contextualist approach to judging teaching against its complex "background" within our "form of life." To exemplify this approach, Stickney presents cases of classroom practice…
Evidence-Based Practice for Children with Speech Sound Disorders: Part 1 Narrative Review
ERIC Educational Resources Information Center
Baker, Elise; McLeod, Sharynne
2011-01-01
Purpose: This article provides a comprehensive narrative review of intervention studies for children with speech sound disorders (SSD). Its companion paper (Baker & McLeod, 2011) provides a tutorial and clinical example of how speech-language pathologists (SLPs) can engage in evidence-based practice (EBP) for this clinical population. Method:…
ERIC Educational Resources Information Center
Lousada, M.; Jesus, Luis M. T.; Hall, A.; Joffe, V.
2014-01-01
Background: The effectiveness of two treatment approaches (phonological therapy and articulation therapy) for treatment of 14 children, aged 4;0-6;7 years, with phonologically based speech-sound disorder (SSD) has been previously analysed with severity outcome measures (percentage of consonants correct score, percentage occurrence of phonological…
NASA Astrophysics Data System (ADS)
Bi, ChuanXing; Jing, WenQian; Zhang, YongBin; Xu, Liang
2015-02-01
The conventional nearfield acoustic holography (NAH) is usually based on the assumption of free-field conditions, and it also requires that the measurement aperture should be larger than the actual source. This paper is to focus on the problem that neither of the above-mentioned requirements can be met, and to examine the feasibility of reconstructing the sound field radiated by partial source, based on double-layer pressure measurements made in a non-free field by using patch NAH combined with sound field separation technique. And also, the sensitivity of the reconstructed result to the measurement error is analyzed in detail. Two experiments involving two speakers in an exterior space and one speaker inside a car cabin are presented. The experimental results demonstrate that the patch NAH based on single-layer pressure measurement cannot obtain a satisfied result due to the influences of disturbing sources and reflections, while the patch NAH based on double-layer pressure measurements can successfully remove these influences and reconstruct the patch sound field effectively.