Sample records for moving sound source

  1. Sound source identification and sound radiation modeling in a moving medium using the time-domain equivalent source method.

    PubMed

    Zhang, Xiao-Zheng; Bi, Chuan-Xing; Zhang, Yong-Bin; Xu, Liang

    2015-05-01

    Planar near-field acoustic holography has been successfully extended to reconstruct the sound field in a moving medium, however, the reconstructed field still contains the convection effect that might lead to the wrong identification of sound sources. In order to accurately identify sound sources in a moving medium, a time-domain equivalent source method is developed. In the method, the real source is replaced by a series of time-domain equivalent sources whose strengths are solved iteratively by utilizing the measured pressure and the known convective time-domain Green's function, and time averaging is used to reduce the instability in the iterative solving process. Since these solved equivalent source strengths are independent of the convection effect, they can be used not only to identify sound sources but also to model sound radiations in both moving and static media. Numerical simulations are performed to investigate the influence of noise on the solved equivalent source strengths and the effect of time averaging on reducing the instability, and to demonstrate the advantages of the proposed method on the source identification and sound radiation modeling.

  2. Wave field synthesis of moving virtual sound sources with complex radiation properties.

    PubMed

    Ahrens, Jens; Spors, Sascha

    2011-11-01

    An approach to the synthesis of moving virtual sound sources with complex radiation properties in wave field synthesis is presented. The approach exploits the fact that any stationary sound source of finite spatial extent radiates spherical waves at sufficient distance. The angular dependency of the radiation properties of the source under consideration is reflected by the amplitude and phase distribution on the spherical wave fronts. The sound field emitted by a uniformly moving monopole source is derived and the far-field radiation properties of the complex virtual source under consideration are incorporated in order to derive a closed-form expression for the loudspeaker driving signal. The results are illustrated via numerical simulations of the synthesis of the sound field of a sample moving complex virtual source.

  3. Binaural Processing of Multiple Sound Sources

    DTIC Science & Technology

    2016-08-18

    Sound Source Localization Identification, and Sound Source Localization When Listeners Move. The CI research was also supported by an NIH grant...8217Cochlear Implant Performance in Realistic Listening Environments,’ Dr. Michael Dorman, Principal Investigator, Dr. William Yost unpaid advisor. The other... Listeners Move. The CI research was also supported by an NIH grant (“Cochlear Implant Performance in Realistic Listening Environments,” Dr. Michael Dorman

  4. Angle-Dependent Distortions in the Perceptual Topology of Acoustic Space

    PubMed Central

    2018-01-01

    By moving sounds around the head and asking listeners to report which ones moved more, it was found that sound sources at the side of a listener must move at least twice as much as ones in front to be judged as moving the same amount. A relative expansion of space in the front and compression at the side has consequences for spatial perception of moving sounds by both static and moving listeners. An accompanying prediction that the apparent location of static sound sources ought to also be distorted agrees with previous work and suggests that this is a general perceptual phenomenon that is not limited to moving signals. A mathematical model that mimics the measured expansion of space can be used to successfully capture several previous findings in spatial auditory perception. The inverse of this function could be used alongside individualized head-related transfer functions and motion tracking to produce hyperstable virtual acoustic environments. PMID:29764312

  5. Wave Field Synthesis of moving sources with arbitrary trajectory and velocity profile.

    PubMed

    Firtha, Gergely; Fiala, Péter

    2017-08-01

    The sound field synthesis of moving sound sources is of great importance when dynamic virtual sound scenes are to be reconstructed. Previous solutions considered only virtual sources moving uniformly along a straight trajectory, synthesized employing a linear loudspeaker array. This article presents the synthesis of point sources following an arbitrary trajectory. Under high-frequency assumptions 2.5D Wave Field Synthesis driving functions are derived for arbitrary shaped secondary source contours by adapting the stationary phase approximation to the dynamic description of sources in motion. It is explained how a referencing function should be chosen in order to optimize the amplitude of synthesis on an arbitrary receiver curve. Finally, a finite difference implementation scheme is considered, making the presented approach suitable for real-time applications.

  6. The effect of spatial distribution on the annoyance caused by simultaneous sounds

    NASA Astrophysics Data System (ADS)

    Vos, Joos; Bronkhorst, Adelbert W.; Fedtke, Thomas

    2004-05-01

    A considerable part of the population is exposed to simultaneous and/or successive environmental sounds from different sources. In many cases, these sources are different with respect to their locations also. In a laboratory study, it was investigated whether the annoyance caused by the multiple sounds is affected by the spatial distribution of the sources. There were four independent variables: (1) sound category (stationary or moving), (2) sound type (stationary: lawn-mower, leaf-blower, and chain saw; moving: road traffic, railway, and motorbike), (3) spatial location (left, right, and combinations), and (4) A-weighted sound exposure level (ASEL of single sources equal to 50, 60, or 70 dB). In addition to the individual sounds in isolation, various combinations of two or three different sources within each sound category and sound level were presented for rating. The annoyance was mainly determined by sound level and sound source type. In most cases there were neither significant main effects of spatial distribution nor significant interaction effects between spatial distribution and the other variables. It was concluded that for rating the spatially distrib- uted sounds investigated, the noise dose can simply be determined by a summation of the levels for the left and right channels. [Work supported by CEU.

  7. The Postural Responses to a Moving Environment of Adults Who Are Blind

    ERIC Educational Resources Information Center

    Stoffregen, Thomas A.; Ito, Kiyohide; Hove, Philip; Yank, Jane Redfield; Bardy, Benoit G.

    2010-01-01

    Adults who are blind stood in a room that could be moved around them. A sound source moved with the room, simulating the acoustic consequences of body sway. Body sway was greater when the room moved than when it was stationary, suggesting that sound may have been used to control stance. (Contains 1 figure.)

  8. Psychophysical evidence for auditory motion parallax.

    PubMed

    Genzel, Daria; Schutte, Michael; Brimijoin, W Owen; MacNeilage, Paul R; Wiegrebe, Lutz

    2018-04-17

    Distance is important: From an ecological perspective, knowledge about the distance to either prey or predator is vital. However, the distance of an unknown sound source is particularly difficult to assess, especially in anechoic environments. In vision, changes in perspective resulting from observer motion produce a reliable, consistent, and unambiguous impression of depth known as motion parallax. Here we demonstrate with formal psychophysics that humans can exploit auditory motion parallax, i.e., the change in the dynamic binaural cues elicited by self-motion, to assess the relative depths of two sound sources. Our data show that sensitivity to relative depth is best when subjects move actively; performance deteriorates when subjects are moved by a motion platform or when the sound sources themselves move. This is true even though the dynamic binaural cues elicited by these three types of motion are identical. Our data demonstrate a perceptual strategy to segregate intermittent sound sources in depth and highlight the tight interaction between self-motion and binaural processing that allows assessment of the spatial layout of complex acoustic scenes.

  9. Sound localization and auditory response capabilities in round goby (Neogobius melanostomus)

    NASA Astrophysics Data System (ADS)

    Rollo, Audrey K.; Higgs, Dennis M.

    2005-04-01

    A fundamental role in vertebrate auditory systems is determining the direction of a sound source. While fish show directional responses to sound, sound localization remains in dispute. The species used in the current study, Neogobius melanostomus (round goby) uses sound in reproductive contexts, with both male and female gobies showing directed movement towards a calling male. The two-choice laboratory experiment was used (active versus quiet speaker) to analyze behavior of gobies in response to sound stimuli. When conspecific male spawning sounds were played, gobies moved in a direct path to the active speaker, suggesting true localization to sound. Of the animals that responded to conspecific sounds, 85% of the females and 66% of the males moved directly to the sound source. Auditory playback of natural and synthetic sounds showed differential behavioral specificity. Of gobies that responded, 89% were attracted to the speaker playing Padogobius martensii sounds, 87% to 100 Hz tone, 62% to white noise, and 56% to Gobius niger sounds. Swimming speed, as well as mean path angle to the speaker, will be presented during the presentation. Results suggest a strong localization of the round goby to a sound source, with some differential sound specificity.

  10. Quantitative measurement of pass-by noise radiated by vehicles running at high speeds

    NASA Astrophysics Data System (ADS)

    Yang, Diange; Wang, Ziteng; Li, Bing; Luo, Yugong; Lian, Xiaomin

    2011-03-01

    It has been a challenge in the past to accurately locate and quantify the pass-by noise source radiated by the running vehicles. A system composed of a microphone array is developed in our current work to do this work. An acoustic-holography method for moving sound sources is designed to handle the Doppler effect effectively in the time domain. The effective sound pressure distribution is reconstructed on the surface of a running vehicle. The method has achieved a high calculation efficiency and is able to quantitatively measure the sound pressure at the sound source and identify the location of the main sound source. The method is also validated by the simulation experiments and the measurement tests with known moving speakers. Finally, the engine noise, tire noise, exhaust noise and wind noise of the vehicle running at different speeds are successfully identified by this method.

  11. Detection of Sound Image Movement During Horizontal Head Rotation

    PubMed Central

    Ohba, Kagesho; Iwaya, Yukio; Suzuki, Yôiti

    2016-01-01

    Movement detection for a virtual sound source was measured during the listener’s horizontal head rotation. Listeners were instructed to do head rotation at a given speed. A trial consisted of two intervals. During an interval, a virtual sound source was presented 60° to the right or left of the listener, who was instructed to rotate the head to face the sound image position. Then in one of a pair of intervals, the sound position was moved slightly in the middle of the rotation. Listeners were asked to judge the interval in a trial during which the sound stimuli moved. Results suggest that detection thresholds are higher when listeners do head rotation. Moreover, this effect was found to be independent of the rotation velocity. PMID:27698993

  12. Active Exhaust Silencing Systen For the Management of Auxillary Power Unit Sound Signatures

    DTIC Science & Technology

    2014-08-01

    conceptual mass-less pistons are introduced into the system before and after the injection site, such that they will move exactly with the plane wave...Unit Sound Signatures, Helminen, et al. Page 2 of 7 either the primary source or the injected source. It is assumed that the pistons are ‘close...source, it causes both pistons to move identically. The pressures induced by the flow on the pistons do not affect the flow generated by the

  13. Issues in Humanoid Audition and Sound Source Localization by Active Audition

    NASA Astrophysics Data System (ADS)

    Nakadai, Kazuhiro; Okuno, Hiroshi G.; Kitano, Hiroaki

    In this paper, we present an active audition system which is implemented on the humanoid robot "SIG the humanoid". The audition system for highly intelligent humanoids localizes sound sources and recognizes auditory events in the auditory scene. Active audition reported in this paper enables SIG to track sources by integrating audition, vision, and motor movements. Given the multiple sound sources in the auditory scene, SIG actively moves its head to improve localization by aligning microphones orthogonal to the sound source and by capturing the possible sound sources by vision. However, such an active head movement inevitably creates motor noises.The system adaptively cancels motor noises using motor control signals and the cover acoustics. The experimental result demonstrates that active audition by integration of audition, vision, and motor control attains sound source tracking in variety of conditions.onditions.

  14. A moving medium formulation for prediction of propeller noise at incidence

    NASA Astrophysics Data System (ADS)

    Ghorbaniasl, Ghader; Lacor, Chris

    2012-01-01

    This paper presents a time domain formulation for the sound field radiated by moving bodies in a uniform steady flow with arbitrary orientation. The aim is to provide a formulation for prediction of noise from body so that effects of crossflow on a propeller can be modeled in the time domain. An established theory of noise generation by a moving source is combined with the moving medium Green's function for derivation of the formulation. A formula with Doppler factor is developed because it is more easily interpreted and is more helpful in examining the physic of systems. Based on the technique presented, the source of asymmetry of the sound field can be explained in terms of physics of a moving source. It is shown that the derived formulation can be interpreted as an extension of formulation 1 and 1A of Farassat based on the Ffowcs Williams and Hawkings (FW-H) equation for moving medium problems. Computational results for a stationary monopole and dipole point source in moving medium, a rotating point force in crossflow, a model of helicopter blade at incidence and a propeller case with subsonic tips at incidence verify the formulation.

  15. Bionic Modeling of Knowledge-Based Guidance in Automated Underwater Vehicles.

    DTIC Science & Technology

    1987-06-24

    bugs and their foraging movements are heard by the sound of rustling leaves or rhythmic wing beats . ASYMMETRY OF EARS The faces of owls have captured...sound source without moving. The barn owl has binaural and monaural cues as well as cues that operate in relative motion when either the target or the...owl moves. Table 1 lists the cues. 7 TM No. 87- 2068 fTable 1. Sound Localization Parameters Used by the Barn Owl I BINAURAL PARAMETERS: 1. the

  16. Egocentric and allocentric representations in auditory cortex

    PubMed Central

    Brimijoin, W. Owen; Bizley, Jennifer K.

    2017-01-01

    A key function of the brain is to provide a stable representation of an object’s location in the world. In hearing, sound azimuth and elevation are encoded by neurons throughout the auditory system, and auditory cortex is necessary for sound localization. However, the coordinate frame in which neurons represent sound space remains undefined: classical spatial receptive fields in head-fixed subjects can be explained either by sensitivity to sound source location relative to the head (egocentric) or relative to the world (allocentric encoding). This coordinate frame ambiguity can be resolved by studying freely moving subjects; here we recorded spatial receptive fields in the auditory cortex of freely moving ferrets. We found that most spatially tuned neurons represented sound source location relative to the head across changes in head position and direction. In addition, we also recorded a small number of neurons in which sound location was represented in a world-centered coordinate frame. We used measurements of spatial tuning across changes in head position and direction to explore the influence of sound source distance and speed of head movement on auditory cortical activity and spatial tuning. Modulation depth of spatial tuning increased with distance for egocentric but not allocentric units, whereas, for both populations, modulation was stronger at faster movement speeds. Our findings suggest that early auditory cortex primarily represents sound source location relative to ourselves but that a minority of cells can represent sound location in the world independent of our own position. PMID:28617796

  17. Analysis of masking effects on speech intelligibility with respect to moving sound stimulus

    NASA Astrophysics Data System (ADS)

    Chen, Chiung Yao

    2004-05-01

    The purpose of this study is to compare the disturbed degree of speech by an immovable noise source and an apparent moving one (AMN). In the study of the sound localization, we found that source-directional sensitivity (SDS) well associates with the magnitude of interaural cross correlation (IACC). Ando et al. [Y. Ando, S. H. Kang, and H. Nagamatsu, J. Acoust. Soc. Jpn. (E) 8, 183-190 (1987)] reported that potential correlation between left and right inferior colliculus at auditory path in the brain is in harmony with the correlation function of amplitude input into two ear-canal entrances. We assume that the degree of disturbance under the apparent moving noisy source is probably different from that being installed in front of us within a constant distance in a free field (no reflection). Then, we found there is a different influence on speech intelligibility between a moving and a fixed source generated by 1/3-octave narrow-band noise with the center frequency 2 kHz. However, the reasons for the moving speed and the masking effects on speech intelligibility were uncertain.

  18. Binaural room simulation

    NASA Technical Reports Server (NTRS)

    Lehnert, H.; Blauert, Jens; Pompetzki, W.

    1991-01-01

    In every-day listening the auditory event perceived by a listener is determined not only by the sound signal that a sound emits but also by a variety of environmental parameters. These parameters are the position, orientation and directional characteristics of the sound source, the listener's position and orientation, the geometrical and acoustical properties of surfaces which affect the sound field and the sound propagation properties of the surrounding fluid. A complete set of these parameters can be called an Acoustic Environment. If the auditory event perceived by a listener is manipulated in such a way that the listener is shifted acoustically into a different acoustic environment without moving himself physically, a Virtual Acoustic Environment has been created. Here, we deal with a special technique to set up nearly arbitrary Virtual Acoustic Environments, the Binaural Room Simulation. The purpose of the Binaural Room Simulation is to compute the binaural impulse response related to a virtual acoustic environment taking into account all parameters mentioned above. One possible way to describe a Virtual Acoustic Environment is the concept of the virtual sound sources. Each of the virtual sources emits a certain signal which is correlated but not necessarily identical with the signal emitted by the direct sound source. If source and receiver are non moving, the acoustic environment becomes a linear time-invariant system. Then, the Binaural Impulse Response from the source to a listener' s eardrums contains all relevant auditory information related to the Virtual Acoustic Environment. Listening into the simulated environment can easily be achieved by convolving the Binaural Impulse Response with dry signals and representing the results via headphones.

  19. Efficient Geometric Sound Propagation Using Visibility Culling

    NASA Astrophysics Data System (ADS)

    Chandak, Anish

    2011-07-01

    Simulating propagation of sound can improve the sense of realism in interactive applications such as video games and can lead to better designs in engineering applications such as architectural acoustics. In this thesis, we present geometric sound propagation techniques which are faster than prior methods and map well to upcoming parallel multi-core CPUs. We model specular reflections by using the image-source method and model finite-edge diffraction by using the well-known Biot-Tolstoy-Medwin (BTM) model. We accelerate the computation of specular reflections by applying novel visibility algorithms, FastV and AD-Frustum, which compute visibility from a point. We accelerate finite-edge diffraction modeling by applying a novel visibility algorithm which computes visibility from a region. Our visibility algorithms are based on frustum tracing and exploit recent advances in fast ray-hierarchy intersections, data-parallel computations, and scalable, multi-core algorithms. The AD-Frustum algorithm adapts its computation to the scene complexity and allows small errors in computing specular reflection paths for higher computational efficiency. FastV and our visibility algorithm from a region are general, object-space, conservative visibility algorithms that together significantly reduce the number of image sources compared to other techniques while preserving the same accuracy. Our geometric propagation algorithms are an order of magnitude faster than prior approaches for modeling specular reflections and two to ten times faster for modeling finite-edge diffraction. Our algorithms are interactive, scale almost linearly on multi-core CPUs, and can handle large, complex, and dynamic scenes. We also compare the accuracy of our sound propagation algorithms with other methods. Once sound propagation is performed, it is desirable to listen to the propagated sound in interactive and engineering applications. We can generate smooth, artifact-free output audio signals by applying efficient audio-processing algorithms. We also present the first efficient audio-processing algorithm for scenarios with simultaneously moving source and moving receiver (MS-MR) which incurs less than 25% overhead compared to static source and moving receiver (SS-MR) or moving source and static receiver (MS-SR) scenario.

  20. Complete de-Dopplerization and acoustic holography for external noise of a high-speed train.

    PubMed

    Yang, Diange; Wen, Junjie; Miao, Feng; Wang, Ziteng; Gu, Xiaoan; Lian, Xiaomin

    2016-09-01

    Identification and measurement of moving sound sources are the bases for vehicle noise control. Acoustic holography has been applied in successfully identifying the moving sound source since the 1990s. However, due to the high demand for the accuracy of holographic data, currently the maximum velocity achieved by acoustic holography is just above 100 km/h. The objective of this study was to establish a method based on the complete Morse acoustic model to restore the measured signal in high-speed situations, and to propose a far-field acoustic holography method applicable for high-speed moving sound sources. Simulated comparisons of the proposed far-field acoustic holography with complete Morse model, the acoustic holography with simplified Morse model and traditional delay-and-sum beamforming were conducted. Experiments with a high-speed train running at the speed of 278 km/h validated the proposed far-field acoustic holography. This study extended the applications of acoustic holography to high-speed situations and established the basis for quantitative measurements of far-field acoustic holography.

  1. Active noise control using a steerable parametric array loudspeaker.

    PubMed

    Tanaka, Nobuo; Tanaka, Motoki

    2010-06-01

    Arguably active noise control enables the sound suppression at the designated control points, while the sound pressure except the targeted locations is likely to augment. The reason is clear; a control source normally radiates the sound omnidirectionally. To cope with this problem, this paper introduces a parametric array loudspeaker (PAL) which produces a spatially focused sound beam due to the attribute of ultrasound used for carrier waves, thereby allowing one to suppress the sound pressure at the designated point without causing spillover in the whole sound field. First the fundamental characteristics of PAL are overviewed. The scattered pressure in the near field contributed by source strength of PAL is then described, which is needed for the design of an active noise control system. Furthermore, the optimal control law for minimizing the sound pressure at control points is derived, the control effect being investigated analytically and experimentally. With a view to tracking a moving target point, a steerable PAL based upon a phased array scheme is presented, with the result that the generation of a moving zone of quiet becomes possible without mechanically rotating the PAL. An experiment is finally conducted, demonstrating the validity of the proposed method.

  2. Techniques and instrumentation for the measurement of transient sound energy flux

    NASA Astrophysics Data System (ADS)

    Watkinson, P. S.; Fahy, F. J.

    1983-12-01

    The evaluation of sound intensity distributions, and sound powers, of essentially continuous sources such as automotive engines, electric motors, production line machinery, furnaces, earth moving machinery and various types of process plants were studied. Although such systems are important sources of community disturbance and, to a lesser extent, of industrial health hazard, the most serious sources of hearing hazard in industry are machines operating on an impact principle, such as drop forges, hammers and punches. Controlled experiments to identify major noise source regions and mechanisms are difficult because it is normally impossible to install them in quiet, anechoic environments. The potential for sound intensity measurement to provide a means of overcoming these difficulties has given promising results, indicating the possibility of separation of directly radiated and reverberant sound fields. However, because of the complexity of transient sound fields, a fundamental investigation is necessary to establish the practicability of intensity field decomposition, which is basic to source characterization techniques.

  3. Photoacoustic effect generated by moving optical sources: Motion in one dimension

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bai, Wenyu; Diebold, Gerald J.

    2016-03-28

    Although the photoacoustic effect is typically generated by pulsed or amplitude modulated optical beams, it is clear from examination of the wave equation for pressure that motion of an optical source in space will result in the production of sound as well. Here, the properties of the photoacoustic effect generated by moving sources in one dimension are investigated. The cases of a moving Gaussian beam, an oscillating delta function source, and an accelerating Gaussian optical sources are reported. The salient feature of one-dimensional sources in the linear acoustic limit is that the amplitude of the beam increases in time withoutmore » bound.« less

  4. Photoacoustic Effect Generated from an Expanding Spherical Source

    NASA Astrophysics Data System (ADS)

    Bai, Wenyu; Diebold, Gerald J.

    2018-02-01

    Although the photoacoustic effect is typically generated by amplitude-modulated continuous or pulsed radiation, the form of the wave equation for pressure that governs the generation of sound indicates that optical sources moving in an absorbing fluid can produce sound as well. Here, the characteristics of the acoustic wave produced by a radially symmetric Gaussian source expanding outwardly from the origin are found. The unique feature of the photoacoustic effect from the spherical source is a trailing compressive wave that arises from reflection of an inwardly propagating component of the wave. Similar to the one-dimensional geometry, an unbounded amplification effect is found for the Gaussian source expanding at the sound speed.

  5. Perceptual constancy in auditory perception of distance to railway tracks.

    PubMed

    De Coensel, Bert; Nilsson, Mats E; Berglund, Birgitta; Brown, A L

    2013-07-01

    Distance to a sound source can be accurately estimated solely from auditory information. With a sound source such as a train that is passing by at a relatively large distance, the most important auditory information for the listener for estimating its distance consists of the intensity of the sound, spectral changes in the sound caused by air absorption, and the motion-induced rate of change of intensity. However, these cues are relative because prior information/experience of the sound source-its source power, its spectrum and the typical speed at which it moves-is required for such distance estimates. This paper describes two listening experiments that allow investigation of further prior contextual information taken into account by listeners-viz., whether they are indoors or outdoors. Asked to estimate the distance to the track of a railway, it is shown that listeners assessing sounds heard inside the dwelling based their distance estimates on the expected train passby sound level outdoors rather than on the passby sound level actually experienced indoors. This form of perceptual constancy may have consequences for the assessment of annoyance caused by railway noise.

  6. Sound Generation in the Presence of Moving Surfaces with Application to Internally Generated Aircraft Engine Noise

    NASA Technical Reports Server (NTRS)

    Goldstein, Marvin E.; Envia, E.

    2002-01-01

    In many cases of technological interest solid boundaries play a direct role in the aerodynamic sound generation process and their presence often results in a large increase in the acoustic radiation. A generalized treatment of the emission of sound from moving boundaries is presented. The approach is similar to that of Ffowcs Williams and Hawkings (1969) but the effect of the surrounding mean flow is explicitly accounted for. The results are used to develop a rational framework for the prediction of internally generated aero-engine noise. The final formulas suggest some new noise sources that may be of practical significance.

  7. Auditory compensation for head rotation is incomplete.

    PubMed

    Freeman, Tom C A; Culling, John F; Akeroyd, Michael A; Brimijoin, W Owen

    2017-02-01

    Hearing is confronted by a similar problem to vision when the observer moves. The image motion that is created remains ambiguous until the observer knows the velocity of eye and/or head. One way the visual system solves this problem is to use motor commands, proprioception, and vestibular information. These "extraretinal signals" compensate for self-movement, converting image motion into head-centered coordinates, although not always perfectly. We investigated whether the auditory system also transforms coordinates by examining the degree of compensation for head rotation when judging a moving sound. Real-time recordings of head motion were used to change the "movement gain" relating head movement to source movement across a loudspeaker array. We then determined psychophysically the gain that corresponded to a perceptually stationary source. Experiment 1 showed that the gain was small and positive for a wide range of trained head speeds. Hence, listeners perceived a stationary source as moving slightly opposite to the head rotation, in much the same way that observers see stationary visual objects move against a smooth pursuit eye movement. Experiment 2 showed the degree of compensation remained the same for sounds presented at different azimuths, although the precision of performance declined when the sound was eccentric. We discuss two possible explanations for incomplete compensation, one based on differences in the accuracy of signals encoding image motion and self-movement and one concerning statistical optimization that sacrifices accuracy for precision. We then consider the degree to which such explanations can be applied to auditory motion perception in moving listeners. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  8. The Coast Artillery Journal. Volume 79, Number 2, March-April 1936

    DTIC Science & Technology

    1936-04-01

    communications, field telephones and wire laying. 4. Tables of Organization and Equipment. For preliminary training, a binaural trainer and an over...with a normal binaural sense and especially acute hearing. After a sufficiently large group of these men has been selected to provide for ex- pected...casualties, they should be given special instruction and practice on the binaural trainer, using a moving sound source. This sound source may be easily

  9. A study of sound generation in subsonic rotors, volume 2

    NASA Technical Reports Server (NTRS)

    Chalupnik, J. D.; Clark, L. T.

    1975-01-01

    Computer programs were developed for use in the analysis of sound generation by subsonic rotors. Program AIRFOIL computes the spectrum of radiated sound from a single airfoil immersed in a laminar flow field. Program ROTOR extends this to a rotating frame, and provides a model for sound generation in subsonic rotors. The program also computes tone sound generation due to steady state forces on the blades. Program TONE uses a moving source analysis to generate a time series for an array of forces moving in a circular path. The resultant time series are than Fourier transformed to render the results in spectral form. Program SDATA is a standard time series analysis package. It reads in two discrete time series and forms auto and cross covariances and normalizes these to form correlations. The program then transforms the covariances to yield auto and cross power spectra by means of a Fourier transformation.

  10. Understanding the Doppler Effect by Analysing Spectrograms of the Sound of a Passing Vehicle

    ERIC Educational Resources Information Center

    Lubyako, Dmitry; Martinez-Piedra, Gordon; Ushenin, Arthur; Ushenin, Arthur; Denvir, Patrick; Dunlop, John; Hall, Alex; Le Roux, Gus; van Someren, Laurence; Weinberger, Harvey

    2017-01-01

    The purpose of this paper is to demonstrate how the Doppler effect can be analysed to deduce information about a moving source of sound waves. Specifically, we find the speed of a car and the distance of its closest approach to an observer using sound recordings from smartphones. A key focus of this paper is how this can be achieved in a…

  11. Levitation of objects using acoustic energy

    NASA Technical Reports Server (NTRS)

    Whymark, R. R.

    1975-01-01

    Activated sound source establishes standing-wave pattern in gap between source and acoustic reflector. Solid or liquid material introduced in region will move to one of the low pressure areas produced at antinodes and remain suspended as long as acoustic signal is present.

  12. Moving Sound Source Localization Based on Sequential Subspace Estimation in Actual Room Environments

    NASA Astrophysics Data System (ADS)

    Tsuji, Daisuke; Suyama, Kenji

    This paper presents a novel method for moving sound source localization and its performance evaluation in actual room environments. The method is based on the MUSIC (MUltiple SIgnal Classification) which is one of the most high resolution localization methods. When using the MUSIC, a computation of eigenvectors of correlation matrix is required for the estimation. It needs often a high computational costs. Especially, in the situation of moving source, it becomes a crucial drawback because the estimation must be conducted at every the observation time. Moreover, since the correlation matrix varies its characteristics due to the spatial-temporal non-stationarity, the matrix have to be estimated using only a few observed samples. It makes the estimation accuracy degraded. In this paper, the PAST (Projection Approximation Subspace Tracking) is applied for sequentially estimating the eigenvectors spanning the subspace. In the PAST, the eigen-decomposition is not required, and therefore it is possible to reduce the computational costs. Several experimental results in the actual room environments are shown to present the superior performance of the proposed method.

  13. Perception of Animacy from the Motion of a Single Sound Object.

    PubMed

    Nielsen, Rasmus Høll; Vuust, Peter; Wallentin, Mikkel

    2015-02-01

    Research in the visual modality has shown that the presence of certain dynamics in the motion of an object has a strong effect on whether or not the entity is perceived as animate. Cues for animacy are, among others, self-propelled motion and direction changes that are seemingly not caused by entities external to, or in direct contact with, the moving object. The present study aimed to extend this research into the auditory domain by determining if similar dynamics could influence the perceived animacy of a sound source. In two experiments, participants were presented with single, synthetically generated 'mosquito' sounds moving along trajectories in space, and asked to rate how certain they were that each sound-emitting entity was alive. At a random point on a linear motion trajectory, the sound source would deviate from its initial path and speed. Results confirm findings from the visual domain that a change in the velocity of motion is positively correlated with perceived animacy, and changes in direction were found to influence animacy judgment as well. This suggests that an ability to facilitate and sustain self-movement is perceived as a living quality not only in the visual domain, but in the auditory domain as well. © 2015 SAGE Publications.

  14. Captive Bottlenose Dolphins Do Discriminate Human-Made Sounds Both Underwater and in the Air.

    PubMed

    Lima, Alice; Sébilleau, Mélissa; Boye, Martin; Durand, Candice; Hausberger, Martine; Lemasson, Alban

    2018-01-01

    Bottlenose dolphins ( Tursiops truncatus ) spontaneously emit individual acoustic signals that identify them to group members. We tested whether these cetaceans could learn artificial individual sound cues played underwater and whether they would generalize this learning to airborne sounds. Dolphins are thought to perceive only underwater sounds and their training depends largely on visual signals. We investigated the behavioral responses of seven dolphins in a group to learned human-made individual sound cues, played underwater and in the air. Dolphins recognized their own sound cue after hearing it underwater as they immediately moved toward the source, whereas when it was airborne they gazed more at the source of their own sound cue but did not approach it. We hypothesize that they perhaps detected modifications of the sound induced by air or were confused by the novelty of the situation, but nevertheless recognized they were being "targeted." They did not respond when hearing another group member's cue in either situation. This study provides further evidence that dolphins respond to individual-specific sounds and that these marine mammals possess some capacity for processing airborne acoustic signals.

  15. Exploring positive hospital ward soundscape interventions.

    PubMed

    Mackrill, J; Jennings, P; Cain, R

    2014-11-01

    Sound is often considered as a negative aspect of an environment that needs mitigating, particularly in hospitals. It is worthwhile however, to consider how subjective responses to hospital sounds can be made more positive. The authors identified natural sound, steady state sound and written sound source information as having the potential to do this. Listening evaluations were conducted with 24 participants who rated their emotional (Relaxation) and cognitive (Interest and Understanding) response to a variety of hospital ward soundscape clips across these three interventions. A repeated measures ANOVA revealed that the 'Relaxation' response was significantly affected (n(2) = 0.05, p = 0.001) by the interventions with natural sound producing a 10.1% more positive response. Most interestingly, written sound source information produced a 4.7% positive change in response. The authors conclude that exploring different ways to improve the sounds of a hospital offers subjective benefits that move beyond sound level reduction. This is an area for future work to focus upon in an effort to achieve more positively experienced hospital soundscapes and environments. Copyright © 2014 Elsevier Ltd and The Ergonomics Society. All rights reserved.

  16. The Contribution of Head Movement to the Externalization and Internalization of Sounds

    PubMed Central

    Brimijoin, W. Owen; Boyd, Alan W.; Akeroyd, Michael A.

    2013-01-01

    Background When stimuli are presented over headphones, they are typically perceived as internalized; i.e., they appear to emanate from inside the head. Sounds presented in the free-field tend to be externalized, i.e., perceived to be emanating from a source in the world. This phenomenon is frequently attributed to reverberation and to the spectral characteristics of the sounds: those sounds whose spectrum and reverberation matches that of free-field signals arriving at the ear canal tend to be more frequently externalized. Another factor, however, is that the virtual location of signals presented over headphones moves in perfect concert with any movements of the head, whereas the location of free-field signals moves in opposition to head movements. The effects of head movement have not been systematically disentangled from reverberation and/or spectral cues, so we measured the degree to which movements contribute to externalization. Methodology/Principal Findings We performed two experiments: 1) Using motion tracking and free-field loudspeaker presentation, we presented signals that moved in their spatial location to match listeners’ head movements. 2) Using motion tracking and binaural room impulse responses, we presented filtered signals over headphones that appeared to remain static relative to the world. The results from experiment 1 showed that free-field signals from the front that move with the head are less likely to be externalized (23%) than those that remain fixed (63%). Experiment 2 showed that virtual signals whose position was fixed relative to the world are more likely to be externalized (65%) than those fixed relative to the head (20%), regardless of the fidelity of the individual impulse responses. Conclusions/Significance Head movements play a significant role in the externalization of sound sources. These findings imply tight integration between binaural cues and self motion cues and underscore the importance of self motion for spatial auditory perception. PMID:24312677

  17. Study of environmental sound source identification based on hidden Markov model for robust speech recognition

    NASA Astrophysics Data System (ADS)

    Nishiura, Takanobu; Nakamura, Satoshi

    2003-10-01

    Humans communicate with each other through speech by focusing on the target speech among environmental sounds in real acoustic environments. We can easily identify the target sound from other environmental sounds. For hands-free speech recognition, the identification of the target speech from environmental sounds is imperative. This mechanism may also be important for a self-moving robot to sense the acoustic environments and communicate with humans. Therefore, this paper first proposes hidden Markov model (HMM)-based environmental sound source identification. Environmental sounds are modeled by three states of HMMs and evaluated using 92 kinds of environmental sounds. The identification accuracy was 95.4%. This paper also proposes a new HMM composition method that composes speech HMMs and an HMM of categorized environmental sounds for robust environmental sound-added speech recognition. As a result of the evaluation experiments, we confirmed that the proposed HMM composition outperforms the conventional HMM composition with speech HMMs and a noise (environmental sound) HMM trained using noise periods prior to the target speech in a captured signal. [Work supported by Ministry of Public Management, Home Affairs, Posts and Telecommunications of Japan.

  18. Generation of Underwater Sound by a Moving High-Power Laser Source.

    DTIC Science & Technology

    1985-08-01

    convolution 63 2. Laser beamwidth effects 67 3. Numerical predictions 69 V Table of Contents (Cont.) Page B. Experimental results 72 1. Pressure...source - II 53 17 Nerfield ef f ects: t as a ftunction of ro and cc 55 19 Nearfield effects : @max as a function of r. and a 56 • 20 Nearfield directivity...of a " photophone " or apparatus for the production of sound by light. Bell’s invention of the photophone was neglected for many years, but recently it

  19. Application of crossed beam technology to direct measurements of sound sources in turbulent jets, part 1

    NASA Technical Reports Server (NTRS)

    Wilson, L. N.

    1970-01-01

    The mathematical bases for the direct measurement of sound source intensities in turbulent jets using the crossed-beam technique are discussed in detail. It is found that the problems associated with such measurements lie in three main areas: (1) measurement of the correct flow covariance, (2) accounting for retarded time effects in the measurements, and (3) transformation of measurements to a moving frame of reference. The determination of the particular conditions under which these problems can be circumvented is the main goal of the study.

  20. The sound of moving bodies. Ph.D. Thesis - Cambridge Univ.

    NASA Technical Reports Server (NTRS)

    Brentner, Kenneth Steven

    1990-01-01

    The importance of the quadrupole source term in the Ffowcs, Williams, and Hawkings (FWH) equation was addressed. The quadrupole source contains fundamental components of the complete fluid mechanics problem, which are ignored only at the risk of error. The results made it clear that any application of the acoustic analogy should begin with all of the source terms in the FWH theory. The direct calculation of the acoustic field as part of the complete unsteady fluid mechanics problem using CFD is considered. It was shown that aeroelastic calculation can indeed be made with CFD codes. The results indicate that the acoustic field is the most susceptible component of the computation to numerical error. Therefore, the ability to measure the damping of acoustic waves is absolutely essential both to develop acoustic computations. Essential groundwork for a new approach to the problem of sound generation by moving bodies is presented. This new computational acoustic approach holds the promise of solving many problems hitherto pushed aside.

  1. Understanding the Doppler effect by analysing spectrograms of the sound of a passing vehicle

    NASA Astrophysics Data System (ADS)

    Lubyako, Dmitry; Martinez-Piedra, Gordon; Ushenin, Arthur; Denvir, Patrick; Dunlop, John; Hall, Alex; Le Roux, Gus; van Someren, Laurence; Weinberger, Harvey

    2017-11-01

    The purpose of this paper is to demonstrate how the Doppler effect can be analysed to deduce information about a moving source of sound waves. Specifically, we find the speed of a car and the distance of its closest approach to an observer using sound recordings from smartphones. A key focus of this paper is how this can be achieved in a classroom, both theoretically and experimentally, to deepen students’ understanding of the Doppler effect. Included are our own experimental data (48 sound recordings) to allow others to reproduce the analysis, if they cannot repeat the whole experiment themselves. In addition to its educational purpose, this paper examines the percentage errors in our results. This enabled us to determine sources of error, allowing those conducting similar future investigations to optimize their accuracy.

  2. Acoustical measurements of sound fields between the stage and the orchestra pit inside an historical opera house

    NASA Astrophysics Data System (ADS)

    Sato, Shin-Ichi; Prodi, Nicola; Sakai, Hiroyuki

    2004-05-01

    To clarify the relationship of the sound fields between the stage and the orchestra pit, we conducted acoustical measurements in a typical historical opera house, the Teatro Comunale of Ferrara, Italy. Orthogonal factors based on the theory of subjective preference and other related factors were analyzed. First, the sound fields for a singer on the stage in relation to the musicians in the pit were analyzed. And then, the sound fields for performers in the pit in relation to the singers on the stage were considered. Because physical factors vary depending on the location of the sound source, performers can move on the stage or in the pit to find the preferred sound field.

  3. Translation of an Object Using Phase-Controlled Sound Sources in Acoustic Levitation

    NASA Astrophysics Data System (ADS)

    Matsui, Takayasu; Ohdaira, Etsuzo; Masuzawa, Nobuyoshi; Ide, Masao

    1995-05-01

    Acoustic levitation is used for positioning materials in the development of new materials in space where there is no gravity. This technique is applicable to materials for which electromagnetic force cannot be used. If the levitation point of the materials can be controlled freely in this application, possibilities of new applications will be extended. In this paper we report on an experimental study on controlling the levitation point of the object in an acoustic levitation system. The system fabricated and tested in this study has two sound sources with vibrating plates facing each other. Translation of the object can be achieved by controlling the phase of the energizing electrical signal for one of the sound sources. It was found that the levitation point can be moved smoothly in proportion to the phase difference between the vibrating plates.

  4. Captive Bottlenose Dolphins Do Discriminate Human-Made Sounds Both Underwater and in the Air

    PubMed Central

    Lima, Alice; Sébilleau, Mélissa; Boye, Martin; Durand, Candice; Hausberger, Martine; Lemasson, Alban

    2018-01-01

    Bottlenose dolphins (Tursiops truncatus) spontaneously emit individual acoustic signals that identify them to group members. We tested whether these cetaceans could learn artificial individual sound cues played underwater and whether they would generalize this learning to airborne sounds. Dolphins are thought to perceive only underwater sounds and their training depends largely on visual signals. We investigated the behavioral responses of seven dolphins in a group to learned human-made individual sound cues, played underwater and in the air. Dolphins recognized their own sound cue after hearing it underwater as they immediately moved toward the source, whereas when it was airborne they gazed more at the source of their own sound cue but did not approach it. We hypothesize that they perhaps detected modifications of the sound induced by air or were confused by the novelty of the situation, but nevertheless recognized they were being “targeted.” They did not respond when hearing another group member’s cue in either situation. This study provides further evidence that dolphins respond to individual-specific sounds and that these marine mammals possess some capacity for processing airborne acoustic signals. PMID:29445350

  5. [Auditory aftereffects with approaching and withdrawing sound sources: dependence on trajectory and domain of presentation of adapting stimuli].

    PubMed

    Malinina, E S; Andreeva, I G

    2013-01-01

    The perceptual peculiarities of sound source withdrawing and approaching and their influence on auditory aftereffects were studied in the free field. The radial movement of the auditory adapting stimuli was imitated by two methods: (1) by oppositely directed simultaneous amplitude change of the wideband signals at two loudspeakers placed at 1.1 and 4.5 m from a listener; (2) by an increase or a decrease of the wideband noise amplitude of the impulses at one of the loudspeakers--whether close or distant. The radial auditory movement of test stimuli was imitated by using the first method of imitation of adapting stimuli movement. Nine listeners estimated the direction of test stimuli movement without adaptation (control) and after adaptation. Adapting stimuli were stationary, slowly moving with sound level variation of 2 dB and rapidly moving with variation of 12 dB. The percentage of "withdrawing" responses was used for psychometric curve construction. Three perceptual phenomena were found. The growing louder effect was shown in control series without adaptation. The effect was characterized by a decrease of the number of "withdrawing" responses and overestimation of test stimuli as approaching. The position-dependent aftereffects were noticed after adaptation to the stationary and slowly moving sound stimuli. The aftereffect was manifested as an increase of the number of "withdrawing" responses and overestimation of test stimuli as withdrawal. The effect was reduced with increase of the distance between the listener and the loudspeaker. Movement aftereffects were revealed after adaptation to the rapidly moving stimuli. Aftereffects were direction-dependent: the number of "withdrawal" responses after adaptation to approach increased, whereas after adaptation to withdrawal it decreased relative to control. The movement aftereffects were more pronounced at imitation of movement of adapting stimuli by the first method. In this case the listener could determine the starting and the finishing points of movement trajectory. Interaction of movement aftereffects with the growing louder effect was absent in all ways of presentation of adapting stimuli. With increase of distance to the source of adapting stimuli, there was observed a tendency for a decrease of aftereffect of approach and for an increase of aftereffect of withdrawal.

  6. Modelling Aerodynamically Generated Sound: Recent Advances in Rotor Noise Prediction

    NASA Technical Reports Server (NTRS)

    Brentner, Kenneth S.

    2000-01-01

    A great deal of progress has been made in the modeling of aerodynamically generated sound for rotors over the past decade. The Ffowcs Williams-Hawkings (FW-H ) equation has been the foundation for much of the development. Both subsonic and supersonic quadrupole noise formulations have been developed for the prediction of high-speed impulsive noise. In an effort to eliminate the need to compute the quadrupole contribution, the FW-H has also been utilized on permeable surfaces surrounding all physical noise sources. Comparison of the Kirchhoff formulation for moving surfaces with the FW-H equation have shown that the Kirchhoff formulation for moving surfaces can give erroneous results for aeroacoustic problems.

  7. Perception by Operators of Approach and Withdrawal of Moving Sound Sources

    DTIC Science & Technology

    1999-01-01

    repeated presentations of signals of the same azimuth value (Figs. 10-11) showed rather great variation in estimations of this subject. In anamnesis of...brain trauma, and in Subj. KL there was an stroke in anamnesis

  8. A Synopsis of Marine Animal Underwater Sounds in Eight Geographic Areas

    DTIC Science & Technology

    1971-05-28

    MAMMALS (1) HIGHEST PROBABILITY Delphinapterus leucas - In Apr.-June they move up Kola River. Present all months except July) Aug, Sept...its vocali- zations are available from Dr. W. C. Cummings, NUC Code 5054. : ii V, , 71 Delphinapterus leucas --(white-whale or beluga) The beluga is...about 5 msoc. ~7 * Delphinapterus leucas --(white whale or beluga-continued Our own estimates of source level for sounds from an individual whale 2 are

  9. Noise from high speed maglev systems: Noise sources, noise criteria, preliminary design guidelines for noise control, recommendations for acoustical test facility for maglev research. Final report, July 1991-October 1992

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hanson, C.E.; Abbot, P.; Dyer, I.

    1993-01-01

    Noise levels from magnetically-levitated trains (maglev) at very high speed may be high enough to cause environmental noise impact in residential areas. Aeroacoustic sources dominate the sound at high speeds and guideway vibrations generate noticeable sound at low speed. In addition to high noise levels, the startle effect as a result of sudden onset of sound from a rapidly moving nearby maglev vehicle may lead to increased annoyance to neighbors of a maglev system. The report provides a base for determining the noise consequences and potential mitigation for a high speed maglev system in populated areas of the United States.more » Four areas are included in the study: (1) definition of noise sources; (2) development of noise criteria; (3) development of design guidelines; and (4) recommendations for a noise testing facility.« less

  10. Object localization using a biosonar beam: how opening your mouth improves localization.

    PubMed

    Arditi, G; Weiss, A J; Yovel, Y

    2015-08-01

    Determining the location of a sound source is crucial for survival. Both predators and prey usually produce sound while moving, revealing valuable information about their presence and location. Animals have thus evolved morphological and neural adaptations allowing precise sound localization. Mammals rely on the temporal and amplitude differences between the sound signals arriving at their two ears, as well as on the spectral cues available in the signal arriving at a single ear to localize a sound source. Most mammals rely on passive hearing and are thus limited by the acoustic characteristics of the emitted sound. Echolocating bats emit sound to perceive their environment. They can, therefore, affect the frequency spectrum of the echoes they must localize. The biosonar sound beam of a bat is directional, spreading different frequencies into different directions. Here, we analyse mathematically the spatial information that is provided by the beam and could be used to improve sound localization. We hypothesize how bats could improve sound localization by altering their echolocation signal design or by increasing their mouth gape (the size of the sound emitter) as they, indeed, do in nature. Finally, we also reveal a trade-off according to which increasing the echolocation signal's frequency improves the accuracy of sound localization but might result in undesired large localization errors under low signal-to-noise ratio conditions.

  11. Object localization using a biosonar beam: how opening your mouth improves localization

    PubMed Central

    Arditi, G.; Weiss, A. J.; Yovel, Y.

    2015-01-01

    Determining the location of a sound source is crucial for survival. Both predators and prey usually produce sound while moving, revealing valuable information about their presence and location. Animals have thus evolved morphological and neural adaptations allowing precise sound localization. Mammals rely on the temporal and amplitude differences between the sound signals arriving at their two ears, as well as on the spectral cues available in the signal arriving at a single ear to localize a sound source. Most mammals rely on passive hearing and are thus limited by the acoustic characteristics of the emitted sound. Echolocating bats emit sound to perceive their environment. They can, therefore, affect the frequency spectrum of the echoes they must localize. The biosonar sound beam of a bat is directional, spreading different frequencies into different directions. Here, we analyse mathematically the spatial information that is provided by the beam and could be used to improve sound localization. We hypothesize how bats could improve sound localization by altering their echolocation signal design or by increasing their mouth gape (the size of the sound emitter) as they, indeed, do in nature. Finally, we also reveal a trade-off according to which increasing the echolocation signal's frequency improves the accuracy of sound localization but might result in undesired large localization errors under low signal-to-noise ratio conditions. PMID:26361552

  12. Behavioral responses of gray whales to industrial noise: feeding observations and predictive modeling

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Malme, C.I.; Wuersig, B.; Bird, J.E.

    1986-08-01

    An investigation was made of the potential effects of underwater noise from petroleum-industry activities on feeding gray whales. The investigation consisted of two components, a field study and an acoustic model study. The field study was performed near Southeast Cape, St. Lawrence Island in August, 1985, using a 100 cu. in. air gun source and playback of drillship noise. Sound-source levels and acoustic-propagation losses were measured to permit estimation of sound exposure levels at whale-sighting positions. For the air-gun source there was a 0.5 probability that the whales would stop feeding and move away from the area when the averagemore » pulse levels reached 173 dB.« less

  13. Locating arbitrarily time-dependent sound sources in three dimensional space in real time.

    PubMed

    Wu, Sean F; Zhu, Na

    2010-08-01

    This paper presents a method for locating arbitrarily time-dependent acoustic sources in a free field in real time by using only four microphones. This method is capable of handling a wide variety of acoustic signals, including broadband, narrowband, impulsive, and continuous sound over the entire audible frequency range, produced by multiple sources in three dimensional (3D) space. Locations of acoustic sources are indicated by the Cartesian coordinates. The underlying principle of this method is a hybrid approach that consists of modeling of acoustic radiation from a point source in a free field, triangulation, and de-noising to enhance the signal to noise ratio (SNR). Numerical simulations are conducted to study the impacts of SNR, microphone spacing, source distance and frequency on spatial resolution and accuracy of source localizations. Based on these results, a simple device that consists of four microphones mounted on three mutually orthogonal axes at an optimal distance, a four-channel signal conditioner, and a camera is fabricated. Experiments are conducted in different environments to assess its effectiveness in locating sources that produce arbitrarily time-dependent acoustic signals, regardless whether a sound source is stationary or moves in space, even toward behind measurement microphones. Practical limitations on this method are discussed.

  14. Statistics of natural binaural sounds.

    PubMed

    Młynarski, Wiktor; Jost, Jürgen

    2014-01-01

    Binaural sound localization is usually considered a discrimination task, where interaural phase (IPD) and level (ILD) disparities at narrowly tuned frequency channels are utilized to identify a position of a sound source. In natural conditions however, binaural circuits are exposed to a stimulation by sound waves originating from multiple, often moving and overlapping sources. Therefore statistics of binaural cues depend on acoustic properties and the spatial configuration of the environment. Distribution of cues encountered naturally and their dependence on physical properties of an auditory scene have not been studied before. In the present work we analyzed statistics of naturally encountered binaural sounds. We performed binaural recordings of three auditory scenes with varying spatial configuration and analyzed empirical cue distributions from each scene. We have found that certain properties such as the spread of IPD distributions as well as an overall shape of ILD distributions do not vary strongly between different auditory scenes. Moreover, we found that ILD distributions vary much weaker across frequency channels and IPDs often attain much higher values, than can be predicted from head filtering properties. In order to understand the complexity of the binaural hearing task in the natural environment, sound waveforms were analyzed by performing Independent Component Analysis (ICA). Properties of learned basis functions indicate that in natural conditions soundwaves in each ear are predominantly generated by independent sources. This implies that the real-world sound localization must rely on mechanisms more complex than a mere cue extraction.

  15. Statistics of Natural Binaural Sounds

    PubMed Central

    Młynarski, Wiktor; Jost, Jürgen

    2014-01-01

    Binaural sound localization is usually considered a discrimination task, where interaural phase (IPD) and level (ILD) disparities at narrowly tuned frequency channels are utilized to identify a position of a sound source. In natural conditions however, binaural circuits are exposed to a stimulation by sound waves originating from multiple, often moving and overlapping sources. Therefore statistics of binaural cues depend on acoustic properties and the spatial configuration of the environment. Distribution of cues encountered naturally and their dependence on physical properties of an auditory scene have not been studied before. In the present work we analyzed statistics of naturally encountered binaural sounds. We performed binaural recordings of three auditory scenes with varying spatial configuration and analyzed empirical cue distributions from each scene. We have found that certain properties such as the spread of IPD distributions as well as an overall shape of ILD distributions do not vary strongly between different auditory scenes. Moreover, we found that ILD distributions vary much weaker across frequency channels and IPDs often attain much higher values, than can be predicted from head filtering properties. In order to understand the complexity of the binaural hearing task in the natural environment, sound waveforms were analyzed by performing Independent Component Analysis (ICA). Properties of learned basis functions indicate that in natural conditions soundwaves in each ear are predominantly generated by independent sources. This implies that the real-world sound localization must rely on mechanisms more complex than a mere cue extraction. PMID:25285658

  16. Short-Latency, Goal-Directed Movements of the Pinnae to Sounds That Produce Auditory Spatial Illusions

    PubMed Central

    McClaine, Elizabeth M.; Yin, Tom C. T.

    2010-01-01

    The precedence effect (PE) is an auditory spatial illusion whereby two identical sounds presented from two separate locations with a delay between them are perceived as a fused single sound source whose position depends on the value of the delay. By training cats using operant conditioning to look at sound sources, we have previously shown that cats experience the PE similarly to humans. For delays less than ±400 μs, cats exhibit summing localization, the perception of a “phantom” sound located between the sources. Consistent with localization dominance, for delays from 400 μs to ∼10 ms, cats orient toward the leading source location only, with little influence of the lagging source. Finally, echo threshold was reached for delays >10 ms, where cats first began to orient to the lagging source. It has been hypothesized by some that the neural mechanisms that produce facets of the PE, such as localization dominance and echo threshold, must likely occur at cortical levels. To test this hypothesis, we measured both pinnae position, which were not under any behavioral constraint, and eye position in cats and found that the pinnae orientations to stimuli that produce each of the three phases of the PE illusion was similar to the gaze responses. Although both eye and pinnae movements behaved in a manner that reflected the PE, because the pinnae moved with strikingly short latencies (∼30 ms), these data suggest a subcortical basis for the PE and that the cortex is not likely to be directly involved. PMID:19889848

  17. Short-latency, goal-directed movements of the pinnae to sounds that produce auditory spatial illusions.

    PubMed

    Tollin, Daniel J; McClaine, Elizabeth M; Yin, Tom C T

    2010-01-01

    The precedence effect (PE) is an auditory spatial illusion whereby two identical sounds presented from two separate locations with a delay between them are perceived as a fused single sound source whose position depends on the value of the delay. By training cats using operant conditioning to look at sound sources, we have previously shown that cats experience the PE similarly to humans. For delays less than +/-400 mus, cats exhibit summing localization, the perception of a "phantom" sound located between the sources. Consistent with localization dominance, for delays from 400 mus to approximately 10 ms, cats orient toward the leading source location only, with little influence of the lagging source. Finally, echo threshold was reached for delays >10 ms, where cats first began to orient to the lagging source. It has been hypothesized by some that the neural mechanisms that produce facets of the PE, such as localization dominance and echo threshold, must likely occur at cortical levels. To test this hypothesis, we measured both pinnae position, which were not under any behavioral constraint, and eye position in cats and found that the pinnae orientations to stimuli that produce each of the three phases of the PE illusion was similar to the gaze responses. Although both eye and pinnae movements behaved in a manner that reflected the PE, because the pinnae moved with strikingly short latencies ( approximately 30 ms), these data suggest a subcortical basis for the PE and that the cortex is not likely to be directly involved.

  18. Modeling effectiveness of gradual increases in source level to mitigate effects of sonar on marine mammals.

    PubMed

    Von Benda-Beckmann, Alexander M; Wensveen, Paul J; Kvadsheim, Petter H; Lam, Frans-Peter A; Miller, Patrick J O; Tyack, Peter L; Ainslie, Michael A

    2014-02-01

    Ramp-up or soft-start procedures (i.e., gradual increase in the source level) are used to mitigate the effect of sonar sound on marine mammals, although no one to date has tested whether ramp-up procedures are effective at reducing the effect of sound on marine mammals. We investigated the effectiveness of ramp-up procedures in reducing the area within which changes in hearing thresholds can occur. We modeled the level of sound killer whales (Orcinus orca) were exposed to from a generic sonar operation preceded by different ramp-up schemes. In our model, ramp-up procedures reduced the risk of killer whales receiving sounds of sufficient intensity to affect their hearing. The effectiveness of the ramp-up procedure depended strongly on the assumed response threshold and differed with ramp-up duration, although extending the duration of the ramp up beyond 5 min did not add much to its predicted mitigating effect. The main factors that limited effectiveness of ramp up in a typical antisubmarine warfare scenario were high source level, rapid moving sonar source, and long silences between consecutive sonar transmissions. Our exposure modeling approach can be used to evaluate and optimize mitigation procedures. © 2013 Society for Conservation Biology.

  19. Photoacoustic trace detection of gases at the parts-per-quadrillion level with a moving optical grating.

    PubMed

    Xiong, Lian; Bai, Wenyu; Chen, Feifei; Zhao, Xian; Yu, Fapeng; Diebold, Gerald J

    2017-07-11

    The amplitude of the photoacoustic effect for an optical source moving at the sound speed in a one-dimensional geometry increases linearly in time without bound in the linear acoustic regime. Here, use of this principle is described for trace detection of gases, using two frequency-shifted beams from a CO 2 laser directed at an angle to each other to give optical fringes that move at the sound speed in a cavity with a longitudinal resonance. The photoacoustic signal is detected with a high-[Formula: see text], piezoelectric crystal with a resonance on the order of [Formula: see text] kHz. The photoacoustic cell has a design analogous to a hemispherical laser resonator and can be adjusted to have a longitudinal resonance to match that of the detector crystal. The grating frequency, the length of the resonator, and the crystal must all have matched frequencies; thus, three resonances are used to advantage to produce sensitivity that extends to the parts-per-quadrillion level.

  20. Effectiveness of Interaural Delays Alone as Cues During Dynamic Sound Localization

    NASA Technical Reports Server (NTRS)

    Wenzel, Elizabeth M.; Null, Cynthia H. (Technical Monitor)

    1996-01-01

    The contribution of interaural time differences (ITDs) to the localization of virtual sound sources with and without head motion was examined. Listeners estimated the apparent azimuth, elevation and distance of virtual sources presented over headphones. Stimuli (3 sec., white noise) were synthesized from minimum-phase representations of nonindividualized head-related transfer functions (HRTFs); binaural magnitude spectra were derived from the minimum phase estimates and ITDs were represented as a pure delay. During dynamic conditions, listeners were encouraged to move their heads; head position was tracked and stimuli were synthesized in real time using a Convolvotron to simulate a stationary external sound source. Two synthesis conditions were tested: (1) both interaural level differences (ILDs) and ITDs correctly correlated with source location and head motion, (2) ITDs correct, no ILDs (flat magnitude spectrum). Head movements reduced azimuth confusions primarily when interaural cues were correctly correlated, although a smaller effect was also seen for ITDs alone. Externalization was generally poor for ITD-only conditions and was enhanced by head motion only for normal HRTFs. Overall the data suggest that, while ITDs alone can provide a significant cue for azimuth, the errors most commonly associated with virtual sources are reduced by location-dependent magnitude cues.

  1. The exact calculation of quadrupole sources for some incompressible flows

    NASA Technical Reports Server (NTRS)

    Brentner, Kenneth S.

    1988-01-01

    This paper is concerned with the application of the acoustic analogy of Lighthill to the acoustic and aerodynamic problems associated with moving bodies. The Ffowcs Williams-Hawkings equation, which is an interpretation of the acoustic analogy for sound generation by moving bodies, manipulates the source terms into surface and volume sources. Quite often in practice the volume sources, or quadrupoles, are neglected for various reasons. Recently, Farassat, Long and others have attempted to use the FW-H equation with the quadrupole source and neglected to solve for the surface pressure on the body. The purpose of this paper is to examine the contribution of the quadrupole source to the acoustic pressure and body surface pressure for some problems for which the exact solution is known. The inviscid, incompressible, 2-D flow, calculated using the velocity potential, is used to calculate the individual contributions of the various surface and volume source terms in the FW-H equation. The relative importance of each of the sources is then assessed.

  2. Auditory Spatial Layout

    NASA Technical Reports Server (NTRS)

    Wightman, Frederic L.; Jenison, Rick

    1995-01-01

    All auditory sensory information is packaged in a pair of acoustical pressure waveforms, one at each ear. While there is obvious structure in these waveforms, that structure (temporal and spectral patterns) bears no simple relationship to the structure of the environmental objects that produced them. The properties of auditory objects and their layout in space must be derived completely from higher level processing of the peripheral input. This chapter begins with a discussion of the peculiarities of acoustical stimuli and how they are received by the human auditory system. A distinction is made between the ambient sound field and the effective stimulus to differentiate the perceptual distinctions among various simple classes of sound sources (ambient field) from the known perceptual consequences of the linear transformations of the sound wave from source to receiver (effective stimulus). Next, the definition of an auditory object is dealt with, specifically the question of how the various components of a sound stream become segregated into distinct auditory objects. The remainder of the chapter focuses on issues related to the spatial layout of auditory objects, both stationary and moving.

  3. 76 FR 47996 - Security Zone; 2011 Seattle Seafair Fleet Week Moving Vessels, Puget Sound, WA; Correction

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-08-08

    ... Zone; 2011 Seattle Seafair Fleet Week Moving Vessels, Puget Sound, WA; Correction AGENCY: Coast Guard...-mail ENS Anthony P. LaBoy, Coast Guard Sector Puget Sound, Waterways Management Division; telephone 206-217- 6323, e-mail SectorPugetSound[email protected] . Correction In the temporary final rule FR Doc. 2011...

  4. Postflight analysis of the single-axis acoustic system on SPAR VI and recommendations for future flights

    NASA Technical Reports Server (NTRS)

    Naumann, R. J.; Oran, W. A.; Whymark, R. R.; Rey, C.

    1981-01-01

    The single axis acoustic levitator that was flown on SPAR VI malfunctioned. The results of a series of tests, analyses, and investigation of hypotheses that were undertaken to determine the probable cause of failure are presented, together with recommendations for future flights of the apparatus. The most probable causes of the SPAR VI failure were lower than expected sound intensity due to mechanical degradation of the sound source, and an unexpected external force that caused the experiment sample to move radially and eventually be lost from the acoustic energy well.

  5. Prediction and reduction of aircraft noise in outdoor environments

    NASA Astrophysics Data System (ADS)

    Tong, Bao N.

    This dissertation investigates the noise due to an en-route aircraft cruising at high altitudes. It offers an improved understanding into the combined effects of atmospheric propagation, ground reflection, and source motion on the impact of en-route aircraft noise. A numerical model has been developed to compute pressure time-histories due to a uniformly moving source above a flat ground surface in the presence of a horizontally stratified atmosphere. For a moving source at high elevations, contributions from a direct and specularly reflected wave are sufficient in predicting the sound field close to the ground. In the absence of wind effects, the predicted sound field from a single overhead flight trajectory can be used to interpolate pressure time histories at all other receiver locations via a simplified ray model for the incoherent sound field. This approach provides an efficient method for generating pressure time histories in a three-dimensional space for noise impact studies. A variety of different noise propagation methods are adapted to a uniformly moving source to evaluate the accuracy and efficiency of their predictions. The techniques include: analytical methods, the Fast Field Program (FFP), and asymptotic analysis methods (e.g., ray tracing and more advanced formulations). Source motion effects are introduced via either a retarded time analysis or a Lorentz transform approach depending on the complexity of the problem. The noise spectrum from a single emission frequency, moving source has broadband characteristics. This is a consequence of the Doppler shift which continuously modifies the perceived frequency of the source as it moves relative to a stationary observer on the ground. Thus, the instantaneous wavefronts must be considered in both the frequency dependent ground impedance model and the atmospheric absorption model. It can be shown that the Doppler factor is invariant along each ray path. This gives rise to a path dependent atmospheric absorption mechanism due to the source's motion. To help mitigate the noise that propagates to the ground, multi-layered acoustic treatments can be applied to provide good performance over a wide range of frequencies. An accurate representation of material properties for each of the constituent layers is needed in the design of such treatments. The parameter of interest is the specific acoustic impedance, which can be obtained via inversion of acoustic transfer function measurements. However, several different impedance values can correspond to the same sound field predictions. The boundary loss factor F (associated with spherical wave reflection) is the source of this ambiguity. A method for identifying the family of solutions and selecting the physically meaningful branch is proposed to resolve this non-uniqueness issue. Accurate deduction of the acoustic impedance depends on precise measurements of the acoustic transfer function. However, measurement uncertainties exists in both the magnitude and the phase of the acoustic transfer function. The ASA/ANSI S1.18 standard impedance deduction method uses phase information, which can be unreliable in many outdoor environments. An improved technique which only relies on magnitude information is developed in this dissertation. A selection of optimal geometries become necessary to reduce the sensitivity of the deduced impedance to small variations in the measured data. A graphical approach is provided which offers greater insight into the optimization problem. A downhill simplex algorithm has been implemented to automate the impedance deduction procedure. Physical constraints are applied to limit the search region and to eliminate rogue solutions. Several case studies consisting of both indoor and outdoor acoustical measurements are presented to validate the proposed technique. The current analysis is limited to locally reacting materials where the acoustic impedance does not depend on the incidence angle of the reflected wave.

  6. A time-domain Kirchhoff formula for the convective acoustic wave equation

    NASA Astrophysics Data System (ADS)

    Ghorbaniasl, Ghader; Siozos-Rousoulis, Leonidas; Lacor, Chris

    2016-03-01

    Kirchhoff's integral method allows propagated sound to be predicted, based on the pressure and its derivatives in time and space obtained on a data surface located in the linear flow region. Kirchhoff's formula for noise prediction from high-speed rotors and propellers suffers from the limitation of the observer located in uniform flow, thus requiring an extension to arbitrarily moving media. This paper presents a Kirchhoff formulation for moving surfaces in a uniform moving medium of arbitrary configuration. First, the convective wave equation is derived in a moving frame, based on the generalized functions theory. The Kirchhoff formula is then obtained for moving surfaces in the time domain. The formula has a similar form to the Kirchhoff formulation for moving surfaces of Farassat and Myers, with the presence of additional terms owing to the moving medium effect. The equation explicitly accounts for the influence of mean flow and angle of attack on the radiated noise. The formula is verified by analytical cases of a monopole source located in a moving medium.

  7. Efficient techniques for wave-based sound propagation in interactive applications

    NASA Astrophysics Data System (ADS)

    Mehra, Ravish

    Sound propagation techniques model the effect of the environment on sound waves and predict their behavior from point of emission at the source to the final point of arrival at the listener. Sound is a pressure wave produced by mechanical vibration of a surface that propagates through a medium such as air or water, and the problem of sound propagation can be formulated mathematically as a second-order partial differential equation called the wave equation. Accurate techniques based on solving the wave equation, also called the wave-based techniques, are too expensive computationally and memory-wise. Therefore, these techniques face many challenges in terms of their applicability in interactive applications including sound propagation in large environments, time-varying source and listener directivity, and high simulation cost for mid-frequencies. In this dissertation, we propose a set of efficient wave-based sound propagation techniques that solve these three challenges and enable the use of wave-based sound propagation in interactive applications. Firstly, we propose a novel equivalent source technique for interactive wave-based sound propagation in large scenes spanning hundreds of meters. It is based on the equivalent source theory used for solving radiation and scattering problems in acoustics and electromagnetics. Instead of using a volumetric or surface-based approach, this technique takes an object-centric approach to sound propagation. The proposed equivalent source technique generates realistic acoustic effects and takes orders of magnitude less runtime memory compared to prior wave-based techniques. Secondly, we present an efficient framework for handling time-varying source and listener directivity for interactive wave-based sound propagation. The source directivity is represented as a linear combination of elementary spherical harmonic sources. This spherical harmonic-based representation of source directivity can support analytical, data-driven, rotating or time-varying directivity function at runtime. Unlike previous approaches, the listener directivity approach can be used to compute spatial audio (3D audio) for a moving, rotating listener at interactive rates. Lastly, we propose an efficient GPU-based time-domain solver for the wave equation that enables wave simulation up to the mid-frequency range in tens of minutes on a desktop computer. It is demonstrated that by carefully mapping all the components of the wave simulator to match the parallel processing capabilities of the graphics processors, significant improvement in performance can be achieved compared to the CPU-based simulators, while maintaining numerical accuracy. We validate these techniques with offline numerical simulations and measured data recorded in an outdoor scene. We present results of preliminary user evaluations conducted to study the impact of these techniques on user's immersion in virtual environment. We have integrated these techniques with the Half-Life 2 game engine, Oculus Rift head-mounted display, and Xbox game controller to enable users to experience high-quality acoustics effects and spatial audio in the virtual environment.

  8. Modeling Aerodynamically Generated Sound of Helicopter Rotors

    NASA Technical Reports Server (NTRS)

    Brentner, Kenneth S.; Farassat, F.

    2002-01-01

    A great deal of progress has been made in the modeling of aerodynamically generated sound of rotors over the past decade. Although the modeling effort has focused on helicopter main rotors, the theory is generally valid for a wide range of rotor configurations. The Ffowcs Williams Hawkings (FW-H) equation has been the foundation for much of the development. The monopole and dipole source terms of the FW-H equation account for the thickness and loading noise, respectively. Bladevortex-interaction noise and broadband noise are important types of loading noise, hence much research has been directed toward the accurate modeling of these noise mechanisms. Both subsonic and supersonic quadrupole noise formulations have been developed for the prediction of high-speed impulsive noise. In an effort to eliminate the need to compute the quadrupole contribution, the FW-H equation has also been utilized on permeable surfaces surrounding all physical noise sources. Comparisons of the Kirchhoff formulation for moving surfaces with the FW-H equation have shown that the Kirchhoff formulation for moving surfaces can give erroneous results for aeroacoustic problems. Finally, significant progress has been made incorporating the rotor noise models into full vehicle noise prediction tools.

  9. Behavioral responses of a harbor porpoise (Phocoena phocoena) to playbacks of broadband pile driving sounds.

    PubMed

    Kastelein, Ronald A; van Heerden, Dorianne; Gransier, Robin; Hoek, Lean

    2013-12-01

    The high under-water sound pressure levels (SPLs) produced during pile driving to build offshore wind turbines may affect harbor porpoises. To estimate the discomfort threshold of pile driving sounds, a porpoise in a quiet pool was exposed to playbacks (46 strikes/min) at five SPLs (6 dB steps: 130-154 dB re 1 μPa). The spectrum of the impulsive sound resembled the spectrum of pile driving sound at tens of kilometers from the pile driving location in shallow water such as that found in the North Sea. The animal's behavior during test and baseline periods was compared. At and above a received broadband SPL of 136 dB re 1 μPa [zero-peak sound pressure level: 151 dB re 1 μPa; t90: 126 ms; sound exposure level of a single strike (SELss): 127 dB re 1 μPa(2) s] the porpoise's respiration rate increased in response to the pile driving sounds. At higher levels, he also jumped out of the water more often. Wild porpoises are expected to move tens of kilometers away from offshore pile driving locations; response distances will vary with context, the sounds' source level, parameters influencing sound propagation, and background noise levels. Copyright © 2013 Elsevier Ltd. All rights reserved.

  10. Refraction of Sound Emitted Near Solid Boundaries from a Sheared Jet

    NASA Technical Reports Server (NTRS)

    Dill, Loren H.; Oyedrian, Ayo A.; Krejsa, Eugene A.

    1998-01-01

    A mathematical model is developed to describe the sound emitted from an arbitrary point within a turbulent flow near solid boundaries. A unidirectional, transversely sheared mean flow is assumed, and the cross-section of the cold jet is of arbitrary shape. The analysis begins with Lilley's formulation of aerodynamic noise and, depending upon the specific model of turbulence used, leads via Fourier analysis to an expression for the spectral density of the intensity of the far-field sound emitted from a unit volume of turbulence. The expressions require solution of a reduced Green's function of Lilley's equation as well as certain moving axis velocity correlations of the turbulence. Integration over the entire flow field is required in order to predict the sound emitted by the complete flow. Calculations are presented for sound emitted from a plugflow jet exiting a semi-infinite flat duct. Polar plots of the far-field directivity show the dependence upon frequency and source position within the duct. Certain model problems are suggested to investigate the effect of duct termination, duct geometry, and mean flow shear upon the far-field sound.

  11. On the temporal window of auditory-brain system in connection with subjective responses

    NASA Astrophysics Data System (ADS)

    Mouri, Kiminori

    2003-08-01

    The human auditory-brain system processes information extracted from autocorrelation function (ACF) of the source signal and interaural cross correlation function (IACF) of binaural sound signals which are associated with the left and right cerebral hemispheres, respectively. The purpose of this dissertation is to determine the desirable temporal window (2T: integration interval) for ACF and IACF mechanisms. For the ACF mechanism, the visual change of Φ(0), i.e., the power of ACF, was associated with the change of loudness, and it is shown that the recommended temporal window is given as about 30(τe)min [s]. The value of (τe)min is the minimum value of effective duration of the running ACF of the source signal. It is worth noticing from the experiment of EEG that the most preferred delay time of the first reflection sound is determined by the piece indicating (τe)min in the source signal. For the IACF mechanism, the temporal window is determined as below: The measured range of τIACC corresponding to subjective angle for the moving image sound depends on the temporal window. Here, the moving image was simulated by the use of two loudspeakers located at +/-20° in the horizontal plane, reproducing amplitude modulated band-limited noise alternatively. It is found that the temporal window has a wide range of values from 0.03 to 1 [s] for the modulation frequency below 0.2 Hz. Thesis advisor: Yoichi Ando Copies of this thesis written in English can be obtained from Kiminori Mouri, 5-3-3-1110 Harayama-dai, Sakai city, Osaka 590-0132, Japan. E-mail address: km529756@aol.com

  12. Minke whale song, spacing, and acoustic communication on the Great Barrier Reef, Australia

    NASA Astrophysics Data System (ADS)

    Gedamke, Jason

    An inquisitive population of minke whale (Balaenoptera acutorostrata ) that concentrates on the Great Barrier Reef during its suspected breeding season offered a unique opportunity to conduct a multi-faceted study of a little-known Balaenopteran species' acoustic behavior. Chapter one investigates whether the minke whale is the source of an unusual, complex, and stereotyped sound recorded, the "star-wars" vocalization. A hydrophone array was towed from a vessel to record sounds from circling whales for subsequent localization of sound sources. These acoustic locations were matched with shipboard and in-water observations of the minke whale, demonstrating the minke whale was the source of this unusual sound. Spectral and temporal features of this sound and the source levels at which it is produced are described. The repetitive "star-wars" vocalization appears similar to the songs of other whale species and has characteristics consistent with reproductive advertisement displays. Chapter two investigates whether song (i.e. the "star-wars" vocalization) has a spacing function through passive monitoring of singer spatial patterns with a moored five-sonobuoy array. Active song playback experiments to singers were also conducted to further test song function. This study demonstrated that singers naturally maintain spatial separations between them through a nearest-neighbor analysis and animated tracks of singer movements. In response to active song playbacks, singers generally moved away and repeated song more quickly suggesting that song repetition interval may help regulate spatial interaction and singer separation. These results further indicate the Great Barrier Reef may be an important reproductive habitat for this species. Chapter three investigates whether song is part of a potentially graded repertoire of acoustic signals. Utilizing both vessel-based recordings and remote recordings from the sonobuoy array, temporal and spectral features, source levels, and associated contextual data of recorded sounds were analyzed. Two categories of sound are described here: (1) patterned song, which was regularly repeated in one of three patterns: slow, fast, and rapid-clustered repetition, and (2) non-patterned "social" sounds recorded from gregarious assemblages of whales. These discrete acoustic signals may comprise a graded system of communication (Slow/fast song → Rapid-clustered song → Social sounds) that is related to the spacing between whales.

  13. Enhanced propagation modeling of directional aviation noise: A hybrid parabolic equation-fast field program method

    NASA Astrophysics Data System (ADS)

    Rosenbaum, Joyce E.

    2011-12-01

    Commercial air traffic is anticipated to increase rapidly in the coming years. The impact of aviation noise on communities surrounding airports is, therefore, a growing concern. Accurate prediction of noise can help to mitigate the impact on communities and foster smoother integration of aerospace engineering advances. The problem of accurate sound level prediction requires careful inclusion of all mechanisms that affect propagation, in addition to correct source characterization. Terrain, ground type, meteorological effects, and source directivity can have a substantial influence on the noise level. Because they are difficult to model, these effects are often included only by rough approximation. This dissertation presents a model designed for sound propagation over uneven terrain, with mixed ground type and realistic meteorological conditions. The model is a hybrid of two numerical techniques: the parabolic equation (PE) and fast field program (FFP) methods, which allow for physics-based inclusion of propagation effects and ensure the low frequency content, a factor in community impact, is predicted accurately. Extension of the hybrid model to a pseudo-three-dimensional representation allows it to produce aviation noise contour maps in the standard form. In order for the model to correctly characterize aviation noise sources, a method of representing arbitrary source directivity patterns was developed for the unique form of the parabolic equation starting field. With this advancement, the model can represent broadband, directional moving sound sources, traveling along user-specified paths. This work was prepared for possible use in the research version of the sound propagation module in the Federal Aviation Administration's new standard predictive tool.

  14. Bone Conduction: Anatomy, Physiology, and Communication

    DTIC Science & Technology

    2007-05-01

    78 7.2 Human Localization Capabilities ..................................................................................84...main functions of the pinna are to direct incoming sound toward the EAC and to aid in sound localization . Some animals (e.g., dogs) can move their...pinnae to aid in sound localization , 9 but humans do not typically have this ability. People who may possess the ability to move their pinnae do

  15. Acoustic Holography

    NASA Astrophysics Data System (ADS)

    Kim, Yang-Hann

    One of the subtle problems that make noise control difficult for engineers is the invisibility of noise or sound. A visual image of noise often helps to determine an appropriate means for noise control. There have been many attempts to fulfill this rather challenging objective. Theoretical (or numerical) means for visualizing the sound field have been attempted, and as a result, a great deal of progress has been made. However, most of these numerical methods are not quite ready for practical applications to noise control problems. In the meantime, rapid progress with instrumentation has made it possible to use multiple microphones and fast signal-processing systems. Although these systems are not perfect, they are useful. A state-of-the-art system has recently become available, but it still has many problematic issues; for example, how can one implement the visualized noise field. The constructed noise or sound picture always consists of bias and random errors, and consequently, it is often difficult to determine the origin of the noise and the spatial distribution of the noise field. Section 26.2 of this chapter introduces a brief history, which is associated with "sound visualization," acoustic source identification methods and what has been accomplished with a line or surface array. Section 26.2.3 introduces difficulties and recent studies, including de-Dopplerization and de-reverberation methods, both essentialfor visualizing a moving noise source, such as occurs for cars or trains. This section also addresses what produces ambiguity in realizing real sound sources in a room or closed space. Another major issue associated with sound/noise visualization is whether or not we can distinguish between mutual dependencies of noise in space (Sect. 26.2.4); for example, we are asked to answer the question, "Can we see two birds singing or one bird with two beaks?"

  16. Acoustic Holography

    NASA Astrophysics Data System (ADS)

    Kim, Yang-Hann

    One of the subtle problems that make noise control difficult for engineers is the invisibility of noise or sound. A visual image of noise often helps to determine an appropriate means for noise control. There have been many attempts to fulfill this rather challenging objective. Theoretical (or numerical) means for visualizing the sound field have been attempted, and as a result, a great deal of progress has been made. However, most of these numerical methods are not quite ready for practical applications to noise control problems. In the meantime, rapid progress with instrumentation has made it possible to use multiple microphones and fast signal-processing systems. Although these systems are not perfect, they are useful. A state-of-the-art system has recently become available, but it still has many problematic issues; for example, how can one implement the visualized noise field. The constructed noise or sound picture always consists of bias and random errors, and consequently, it is often difficult to determine the origin of the noise and the spatial distribution of the noise field. Section 26.2 of this chapter introduces a brief history, which is associated with sound visualization, acoustic source identification methods and what has been accomplished with a line or surface array. Section 26.2.3 introduces difficulties and recent studies, including de-Dopplerization and de-re verberation methods, both essential for visualizing a moving noise source, such as occurs for cars or trains. This section also addresses what produces ambiguity in realizing real sound sources in a room or closed space. Another major issue associated with sound/noise visualization is whether or not we can distinguish between mutual dependencies of noise in space (Sect. 26.2.4); for example, we are asked to answer the question, Can we see two birds singing or one bird with two beaks?

  17. Modulation frequency as a cue for auditory speed perception.

    PubMed

    Senna, Irene; Parise, Cesare V; Ernst, Marc O

    2017-07-12

    Unlike vision, the mechanisms underlying auditory motion perception are poorly understood. Here we describe an auditory motion illusion revealing a novel cue to auditory speed perception: the temporal frequency of amplitude modulation (AM-frequency), typical for rattling sounds. Naturally, corrugated objects sliding across each other generate rattling sounds whose AM-frequency tends to directly correlate with speed. We found that AM-frequency modulates auditory speed perception in a highly systematic fashion: moving sounds with higher AM-frequency are perceived as moving faster than sounds with lower AM-frequency. Even more interestingly, sounds with higher AM-frequency also induce stronger motion aftereffects. This reveals the existence of specialized neural mechanisms for auditory motion perception, which are sensitive to AM-frequency. Thus, in spatial hearing, the brain successfully capitalizes on the AM-frequency of rattling sounds to estimate the speed of moving objects. This tightly parallels previous findings in motion vision, where spatio-temporal frequency of moving displays systematically affects both speed perception and the magnitude of the motion aftereffects. Such an analogy with vision suggests that motion detection may rely on canonical computations, with similar neural mechanisms shared across the different modalities. © 2017 The Author(s).

  18. Discrimination of sound source velocity in human listeners

    NASA Astrophysics Data System (ADS)

    Carlile, Simon; Best, Virginia

    2002-02-01

    The ability of six human subjects to discriminate the velocity of moving sound sources was examined using broadband stimuli presented in virtual auditory space. Subjects were presented with two successive stimuli moving in the frontal horizontal plane level with the ears, and were required to judge which moved the fastest. Discrimination thresholds were calculated for reference velocities of 15, 30, and 60 degrees/s under three stimulus conditions. In one condition, stimuli were centered on 0° azimuth and their duration varied randomly to prevent subjects from using displacement as an indicator of velocity. Performance varied between subjects giving median thresholds of 5.5, 9.1, and 14.8 degrees/s for the three reference velocities, respectively. In a second condition, pairs of stimuli were presented for a constant duration and subjects would have been able to use displacement to assist their judgment as faster stimuli traveled further. It was found that thresholds decreased significantly for all velocities (3.8, 7.1, and 9.8 degrees/s), suggesting that the subjects were using the additional displacement cue. The third condition differed from the second in that the stimuli were ``anchored'' on the same starting location rather than centered on the midline, thus doubling the spatial offset between stimulus endpoints. Subjects showed the lowest thresholds in this condition (2.9, 4.0, and 7.0 degrees/s). The results suggested that the auditory system is sensitive to velocity per se, but velocity comparisons are greatly aided if displacement cues are present.

  19. RICH: OPEN-SOURCE HYDRODYNAMIC SIMULATION ON A MOVING VORONOI MESH

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yalinewich, Almog; Steinberg, Elad; Sari, Re’em

    2015-02-01

    We present here RICH, a state-of-the-art two-dimensional hydrodynamic code based on Godunov’s method, on an unstructured moving mesh (the acronym stands for Racah Institute Computational Hydrodynamics). This code is largely based on the code AREPO. It differs from AREPO in the interpolation and time-advancement schemeS as well as a novel parallelization scheme based on Voronoi tessellation. Using our code, we study the pros and cons of a moving mesh (in comparison to a static mesh). We also compare its accuracy to other codes. Specifically, we show that our implementation of external sources and time-advancement scheme is more accurate and robustmore » than is AREPO when the mesh is allowed to move. We performed a parameter study of the cell rounding mechanism (Lloyd iterations) and its effects. We find that in most cases a moving mesh gives better results than a static mesh, but it is not universally true. In the case where matter moves in one way and a sound wave is traveling in the other way (such that relative to the grid the wave is not moving) a static mesh gives better results than a moving mesh. We perform an analytic analysis for finite difference schemes that reveals that a Lagrangian simulation is better than a Eulerian simulation in the case of a highly supersonic flow. Moreover, we show that Voronoi-based moving mesh schemes suffer from an error, which is resolution independent, due to inconsistencies between the flux calculation and the change in the area of a cell. Our code is publicly available as open source and designed in an object-oriented, user-friendly way that facilitates incorporation of new algorithms and physical processes.« less

  20. A time-domain Kirchhoff formula for the convective acoustic wave equation

    PubMed Central

    Ghorbaniasl, Ghader; Siozos-Rousoulis, Leonidas; Lacor, Chris

    2016-01-01

    Kirchhoff’s integral method allows propagated sound to be predicted, based on the pressure and its derivatives in time and space obtained on a data surface located in the linear flow region. Kirchhoff’s formula for noise prediction from high-speed rotors and propellers suffers from the limitation of the observer located in uniform flow, thus requiring an extension to arbitrarily moving media. This paper presents a Kirchhoff formulation for moving surfaces in a uniform moving medium of arbitrary configuration. First, the convective wave equation is derived in a moving frame, based on the generalized functions theory. The Kirchhoff formula is then obtained for moving surfaces in the time domain. The formula has a similar form to the Kirchhoff formulation for moving surfaces of Farassat and Myers, with the presence of additional terms owing to the moving medium effect. The equation explicitly accounts for the influence of mean flow and angle of attack on the radiated noise. The formula is verified by analytical cases of a monopole source located in a moving medium. PMID:27118912

  1. Intercepting a sound without vision

    PubMed Central

    Vercillo, Tiziana; Tonelli, Alessia; Gori, Monica

    2017-01-01

    Visual information is extremely important to generate internal spatial representations. In the auditory modality, the absence of visual cues during early infancy does not preclude the development of some spatial strategies. However, specific spatial abilities might result impaired. In the current study, we investigated the effect of early visual deprivation on the ability to localize static and moving auditory stimuli by comparing sighted and early blind individuals’ performance in different spatial tasks. We also examined perceptual stability in the two groups of participants by matching localization accuracy in a static and a dynamic head condition that involved rotational head movements. Sighted participants accurately localized static and moving sounds. Their localization ability remained unchanged after rotational movements of the head. Conversely, blind participants showed a leftward bias during the localization of static sounds and a little bias for moving sounds. Moreover, head movements induced a significant bias in the direction of head motion during the localization of moving sounds. These results suggest that internal spatial representations might be body-centered in blind individuals and that in sighted people the availability of visual cues during early infancy may affect sensory-motor interactions. PMID:28481939

  2. 77 FR 44475 - Security Zones; Seattle's Seafair Fleet Week Moving Vessels, Puget Sound, WA

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-07-30

    ...-AA87 Security Zones; Seattle's Seafair Fleet Week Moving Vessels, Puget Sound, WA AGENCY: Coast Guard... Seafair Fleet Week Moving Vessels Security Zones from 12:00 p.m. on July 31, 2012 through 5:00 p.m. on August 6, 2012. These security zones are necessary to help ensure the security of the vessels from...

  3. Angry Birds in Space

    NASA Astrophysics Data System (ADS)

    Halford, A. J.

    2017-12-01

    When space computers first started listening into space radio, they noticed that there were radio noises that happened on the morning side of the Earth. Because these waves sounded like noises birds make in the morning, we named these waves after them. These bird sounding waves can move around the Earth, flying up and down, and sometimes move into an area where there is more stuff. This area is also much colder than where these bird noises are first made. When the waves move into this cold area where there is more stuff, they start to sound like angry birds instead of happy birds. Both of these waves, the happy and angry bird sounding waves, are very important to our understanding of how the tiny things in space move and change. Sometimes the waves which sound like birds can push these tiniest of things into the sky. The happy bird sounding waves can push the tiniest things quickly while the angry bird sounding waves push the tinest of things more slowly. When the tiny things fall into the sky, they create beautiful space lights and light that burns which can hurt people in up goers and not so up goers as well as our things like phones, and space computers. We study these waves that sound like birds to better understand when and where the tiny things will fall. That way we can be prepared and enjoy watching the pretty space lights at night with no worries.

  4. A Framework for Simulation of Aircraft Flyover Noise Through a Non-Standard Atmosphere

    NASA Technical Reports Server (NTRS)

    Arntzen, Michael; Rizzi, Stephen A.; Visser, Hendrikus G.; Simons, Dick G.

    2012-01-01

    This paper describes a new framework for the simulation of aircraft flyover noise through a non-standard atmosphere. Central to the framework is a ray-tracing algorithm which defines multiple curved propagation paths, if the atmosphere allows, between the moving source and listener. Because each path has a different emission angle, synthesis of the sound at the source must be performed independently for each path. The time delay, spreading loss and absorption (ground and atmosphere) are integrated along each path, and applied to each synthesized aircraft noise source to simulate a flyover. A final step assigns each resulting signal to its corresponding receiver angle for the simulation of a flyover in a virtual reality environment. Spectrograms of the results from a straight path and a curved path modeling assumption are shown. When the aircraft is at close range, the straight path results are valid. Differences appear especially when the source is relatively far away at shallow elevation angles. These differences, however, are not significant in common sound metrics. While the framework used in this work performs off-line processing, it is conducive to real-time implementation.

  5. The Ffowcs Williams-Hawkings equation - Fifteen years of research

    NASA Technical Reports Server (NTRS)

    Farassat, F.

    1986-01-01

    The Ffowcs Williams-Hawkings equation governs the generation of sound in fluids in the presence of solid boundaries in motion. This equation is reviewed for situations where the linearization of the governing equations is allowed. In addition, research on the application of this equation to problems of aeroacoustic is briefly surveyed. Particular attention is given to the formulation of supersonic sources moving in uniform propeller-like motion.

  6. Sound Clocks and Sonic Relativity

    NASA Astrophysics Data System (ADS)

    Todd, Scott L.; Menicucci, Nicolas C.

    2017-10-01

    Sound propagation within certain non-relativistic condensed matter models obeys a relativistic wave equation despite such systems admitting entirely non-relativistic descriptions. A natural question that arises upon consideration of this is, "do devices exist that will experience the relativity in these systems?" We describe a thought experiment in which `acoustic observers' possess devices called sound clocks that can be connected to form chains. Careful investigation shows that appropriately constructed chains of stationary and moving sound clocks are perceived by observers on the other chain as undergoing the relativistic phenomena of length contraction and time dilation by the Lorentz factor, γ , with c the speed of sound. Sound clocks within moving chains actually tick less frequently than stationary ones and must be separated by a shorter distance than when stationary to satisfy simultaneity conditions. Stationary sound clocks appear to be length contracted and time dilated to moving observers due to their misunderstanding of their own state of motion with respect to the laboratory. Observers restricted to using sound clocks describe a universe kinematically consistent with the theory of special relativity, despite the preferred frame of their universe in the laboratory. Such devices show promise in further probing analogue relativity models, for example in investigating phenomena that require careful consideration of the proper time elapsed for observers.

  7. Detection and tracking of drones using advanced acoustic cameras

    NASA Astrophysics Data System (ADS)

    Busset, Joël.; Perrodin, Florian; Wellig, Peter; Ott, Beat; Heutschi, Kurt; Rühl, Torben; Nussbaumer, Thomas

    2015-10-01

    Recent events of drones flying over city centers, official buildings and nuclear installations stressed the growing threat of uncontrolled drone proliferation and the lack of real countermeasure. Indeed, detecting and tracking them can be difficult with traditional techniques. A system to acoustically detect and track small moving objects, such as drones or ground robots, using acoustic cameras is presented. The described sensor, is completely passive, and composed of a 120-element microphone array and a video camera. The acoustic imaging algorithm determines in real-time the sound power level coming from all directions, using the phase of the sound signals. A tracking algorithm is then able to follow the sound sources. Additionally, a beamforming algorithm selectively extracts the sound coming from each tracked sound source. This extracted sound signal can be used to identify sound signatures and determine the type of object. The described techniques can detect and track any object that produces noise (engines, propellers, tires, etc). It is a good complementary approach to more traditional techniques such as (i) optical and infrared cameras, for which the object may only represent few pixels and may be hidden by the blooming of a bright background, and (ii) radar or other echo-localization techniques, suffering from the weakness of the echo signal coming back to the sensor. The distance of detection depends on the type (frequency range) and volume of the noise emitted by the object, and on the background noise of the environment. Detection range and resilience to background noise were tested in both, laboratory environments and outdoor conditions. It was determined that drones can be tracked up to 160 to 250 meters, depending on their type. Speech extraction was also experimentally investigated: the speech signal of a person being 80 to 100 meters away can be captured with acceptable speech intelligibility.

  8. Seasonal and ontogenetic changes in movement patterns of sixgill sharks.

    PubMed

    Andrews, Kelly S; Williams, Greg D; Levin, Phillip S

    2010-09-08

    Understanding movement patterns is fundamental to population and conservation biology. The way an animal moves through its environment influences the dynamics of local populations and will determine how susceptible it is to natural or anthropogenic perturbations. It is of particular interest to understand the patterns of movement for species which are susceptible to human activities (e.g. fishing), or that exert a large influence on community structure, such as sharks. We monitored the patterns of movement of 34 sixgill sharks Hexanchus griseus using two large-scale acoustic arrays inside and outside Puget Sound, Washington, USA. Sixgill sharks were residents in Puget Sound for up to at least four years before making large movements out of the estuary. Within Puget Sound, sixgills inhabited sites for several weeks at a time and returned to the same sites annually. Across four years, sixgills had consistent seasonal movements in which they moved to the north from winter to spring and moved to the south from summer to fall. Just prior to leaving Puget Sound, sixgills altered their behavior and moved twice as fast among sites. Nineteen of the thirty-four sixgills were detected leaving Puget Sound for the outer coast. Three of these sharks returned to Puget Sound. For most large marine predators, we have a limited understanding of how they move through their environment, and this clouds our ability to successfully manage their populations and their communities. With detailed movement information, such as that being uncovered with acoustic monitoring, we can begin to quantify the spatial and temporal impacts of large predators within the framework of their ecosystems.

  9. Patterns of Macrozooplankton and Fish Occurrence Beneath McMurdo Sound Fast Ice during Spring/Summer 2014/2015

    NASA Astrophysics Data System (ADS)

    Saenz, B. T.; Daly, K. L.; Kim, S.; Ainley, D. G.; Ballard, G.

    2016-02-01

    McMurdo Sound, Antarctica, represents a unique environment for study of trophic interactions, where a full complement of marine predators thrive. As part of a greater study of McMurdo Sound food web interactions, including ocean and ice physics, algal characterization, and predator behavior, macrozooplankton and fish were surveyed using bioacoustics and video using a specially-designed under-ice ROV. Acoustic returns from 82 under-ice surveys were divided into classes consisting of krill, silverfish, and weak scatters. Krill were scarce during surveys in late November, but increased in abundance in association with increasing chlorophyll a in December and early January when surveys ended. The greatest concentrations of krill were found near Ross Island in the eastern Sound, where southerly currents move high-productivity waters beneath the fast ice. Conversely, silverfish, especially schools of juveniles, were found in greater abundance toward the west where currents flow northward and platelet ice typically blocked light from surface waters. Silverfish were rare toward the end of the survey in late December/early January, but possibly had moved deeper than the acoustic instrument could detect. Overall, krill were less abundant and occurred deeper in the water column within 2 km of the fast ice edge, which was accessible by air-breathing predators, suggesting that predation pressure helped structure krill abundance or distribution. Acoustic returns from weak scatters, which included observed jellies, pteropods, detached ice algae and potentially other mesoplankton in high abundance such as copepods, also increased during the study period and co-occurred with chlorophyll a. The patterns of macrozooplankton and fish observed in McMurdo Sound raise important questions about source-sink dynamics, overwinter strategies of mid-trophic organisms, prey-predator dynamics, and sea-ice structuring of ecosystems.

  10. Interactive Sound Propagation using Precomputation and Statistical Approximations

    NASA Astrophysics Data System (ADS)

    Antani, Lakulish

    Acoustic phenomena such as early reflections, diffraction, and reverberation have been shown to improve the user experience in interactive virtual environments and video games. These effects arise due to repeated interactions between sound waves and objects in the environment. In interactive applications, these effects must be simulated within a prescribed time budget. We present two complementary approaches for computing such acoustic effects in real time, with plausible variation in the sound field throughout the scene. The first approach, Precomputed Acoustic Radiance Transfer, precomputes a matrix that accounts for multiple acoustic interactions between all scene objects. The matrix is used at run time to provide sound propagation effects that vary smoothly as sources and listeners move. The second approach couples two techniques---Ambient Reverberance, and Aural Proxies---to provide approximate sound propagation effects in real time, based on only the portion of the environment immediately visible to the listener. These approaches lie at different ends of a space of interactive sound propagation techniques for modeling sound propagation effects in interactive applications. The first approach emphasizes accuracy by modeling acoustic interactions between all parts of the scene; the second approach emphasizes efficiency by only taking the local environment of the listener into account. These methods have been used to efficiently generate acoustic walkthroughs of architectural models. They have also been integrated into a modern game engine, and can enable realistic, interactive sound propagation on commodity desktop PCs.

  11. Interactive physically-based sound simulation

    NASA Astrophysics Data System (ADS)

    Raghuvanshi, Nikunj

    The realization of interactive, immersive virtual worlds requires the ability to present a realistic audio experience that convincingly compliments their visual rendering. Physical simulation is a natural way to achieve such realism, enabling deeply immersive virtual worlds. However, physically-based sound simulation is very computationally expensive owing to the high-frequency, transient oscillations underlying audible sounds. The increasing computational power of desktop computers has served to reduce the gap between required and available computation, and it has become possible to bridge this gap further by using a combination of algorithmic improvements that exploit the physical, as well as perceptual properties of audible sounds. My thesis is a step in this direction. My dissertation concentrates on developing real-time techniques for both sub-problems of sound simulation: synthesis and propagation. Sound synthesis is concerned with generating the sounds produced by objects due to elastic surface vibrations upon interaction with the environment, such as collisions. I present novel techniques that exploit human auditory perception to simulate scenes with hundreds of sounding objects undergoing impact and rolling in real time. Sound propagation is the complementary problem of modeling the high-order scattering and diffraction of sound in an environment as it travels from source to listener. I discuss my work on a novel numerical acoustic simulator (ARD) that is hundred times faster and consumes ten times less memory than a high-accuracy finite-difference technique, allowing acoustic simulations on previously-intractable spaces, such as a cathedral, on a desktop computer. Lastly, I present my work on interactive sound propagation that leverages my ARD simulator to render the acoustics of arbitrary static scenes for multiple moving sources and listener in real time, while accounting for scene-dependent effects such as low-pass filtering and smooth attenuation behind obstructions, reverberation, scattering from complex geometry and sound focusing. This is enabled by a novel compact representation that takes a thousand times less memory than a direct scheme, thus reducing memory footprints to fit within available main memory. To the best of my knowledge, this is the only technique and system in existence to demonstrate auralization of physical wave-based effects in real-time on large, complex 3D scenes.

  12. Vector Acoustics, Vector Sensors, and 3D Underwater Imaging

    NASA Astrophysics Data System (ADS)

    Lindwall, D.

    2007-12-01

    Vector acoustic data has two more dimensions of information than pressure data and may allow for 3D underwater imaging with much less data than with hydrophone data. The vector acoustic sensors measures the particle motions due to passing sound waves and, in conjunction with a collocated hydrophone, the direction of travel of the sound waves. When using a controlled source with known source and sensor locations, the reflection points of the sound field can be determined with a simple trigonometric calculation. I demonstrate this concept with an experiment that used an accelerometer based vector acoustic sensor in a water tank with a short-pulse source and passive scattering targets. The sensor consists of a three-axis accelerometer and a matched hydrophone. The sound source was a standard transducer driven by a short 7 kHz pulse. The sensor was suspended in a fixed location and the hydrophone was moved about the tank by a robotic arm to insonify the tank from many locations. Several floats were placed in the tank as acoustic targets at diagonal ranges of approximately one meter. The accelerometer data show the direct source wave as well as the target scattered waves and reflections from the nearby water surface, tank bottom and sides. Without resorting to the usual methods of seismic imaging, which in this case is only two dimensional and relied entirely on the use of a synthetic source aperture, the two targets, the tank walls, the tank bottom, and the water surface were imaged. A directional ambiguity inherent to vector sensors is removed by using collocated hydrophone data. Although this experiment was in a very simple environment, it suggests that 3-D seismic surveys may be achieved with vector sensors using the same logistics as a 2-D survey that uses conventional hydrophones. This work was supported by the Office of Naval Research, program element 61153N.

  13. 77 FR 40521 - Security Zones, Seattle's Seafair Fleet Week Moving Vessels, Puget Sound, WA

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-07-10

    ... 1625-AA87 Security Zones, Seattle's Seafair Fleet Week Moving Vessels, Puget Sound, WA AGENCY: Coast Guard, DHS. ACTION: Final rule. SUMMARY: The U.S. Coast Guard is establishing security zones around designated participating vessels that are not protected by the Naval Vessel Protection Zone in Seattle's...

  14. Compressional and Shear Wakes in a 2D Dusty Plasma Crystal

    NASA Astrophysics Data System (ADS)

    Nosenko, V.; Goree, J.; Ma, Z. W.; Dubin, D. H. E.

    2001-10-01

    A 2D crystalline lattice can vibrate with two kinds of sound waves, compressional and shear (transverse), where the latter has a much slower sound speed. When these waves are excited by a moving supersonic disturbance, the superposition of the waves creates a Mach cone, i.e., a V-shaped wake. In our experiments, the supersonic disturbance was a moving spot of argon laser light, and this laser light applied a force, due to radiation pressure, on the particles. The beam was swept across the lattice in a controlled and repeatable manner. The particles were levitated in an argon rf discharge. By moving the laser spot faster than the shear sound speed c_t, but slower than the compressional sound speed c_l, we excited a shear wave Mach cone. Alternatively, by moving the laser spot faster than c_l, we excited both cones. In addition to Mach cones, we also observed a wake structure that arises from the compressional wave’s dispersion. We compare our results to Dubin’s theory (Phys. Plasmas 2000) and to molecular dynamics (MD) simulations.

  15. Development of a noncompact source theory with applications to helicopter rotors

    NASA Technical Reports Server (NTRS)

    Farassat, F.; Brown, T. J.

    1976-01-01

    A new formulation for determining the acoustic field of moving bodies, based on acoustic analogy, is derived. The acoustic pressure is given as the sum of two integrals, one of which has a derivative with respect to time. The integrands are functions of the normal velocity and surface pressure of the body. A computer program based on this formulation was used to calculate acoustic pressure signatures for several helicoptor rotors from experimental surface pressure data. Results are compared with those from compact source calculations. It is shown that noncompactness of steady sources on the rotor can account for the high harmonics of the pressure system. Thickness noise is shown to be a significant source of sound, especially for blunt airfoils in regions where noncompact source theory should be applied.

  16. Making Sound Connections

    ERIC Educational Resources Information Center

    Deal, Walter F., III

    2007-01-01

    Sound provides and offers amazing insights into the world. Sound waves may be defined as mechanical energy that moves through air or other medium as a longitudinal wave and consists of pressure fluctuations. Humans and animals alike use sound as a means of communication and a tool for survival. Mammals, such as bats, use ultrasonic sound waves to…

  17. Seasonal and Ontogenetic Changes in Movement Patterns of Sixgill Sharks

    PubMed Central

    Andrews, Kelly S.; Williams, Greg D.; Levin, Phillip S.

    2010-01-01

    Background Understanding movement patterns is fundamental to population and conservation biology. The way an animal moves through its environment influences the dynamics of local populations and will determine how susceptible it is to natural or anthropogenic perturbations. It is of particular interest to understand the patterns of movement for species which are susceptible to human activities (e.g. fishing), or that exert a large influence on community structure, such as sharks. Methodology/Principal Findings We monitored the patterns of movement of 34 sixgill sharks Hexanchus griseus using two large-scale acoustic arrays inside and outside Puget Sound, Washington, USA. Sixgill sharks were residents in Puget Sound for up to at least four years before making large movements out of the estuary. Within Puget Sound, sixgills inhabited sites for several weeks at a time and returned to the same sites annually. Across four years, sixgills had consistent seasonal movements in which they moved to the north from winter to spring and moved to the south from summer to fall. Just prior to leaving Puget Sound, sixgills altered their behavior and moved twice as fast among sites. Nineteen of the thirty-four sixgills were detected leaving Puget Sound for the outer coast. Three of these sharks returned to Puget Sound. Conclusions/Significance For most large marine predators, we have a limited understanding of how they move through their environment, and this clouds our ability to successfully manage their populations and their communities. With detailed movement information, such as that being uncovered with acoustic monitoring, we can begin to quantify the spatial and temporal impacts of large predators within the framework of their ecosystems. PMID:20838617

  18. Dependence of auditory spatial updating on vestibular, proprioceptive, and efference copy signals

    PubMed Central

    Genzel, Daria; Firzlaff, Uwe; Wiegrebe, Lutz

    2016-01-01

    Humans localize sounds by comparing inputs across the two ears, resulting in a head-centered representation of sound-source position. When the head moves, information about head movement must be combined with the head-centered estimate to correctly update the world-centered sound-source position. Spatial updating has been extensively studied in the visual system, but less is known about how head movement signals interact with binaural information during auditory spatial updating. In the current experiments, listeners compared the world-centered azimuthal position of two sound sources presented before and after a head rotation that depended on condition. In the active condition, subjects rotated their head by ∼35° to the left or right, following a pretrained trajectory. In the passive condition, subjects were rotated along the same trajectory in a rotating chair. In the cancellation condition, subjects rotated their head as in the active condition, but the chair was counter-rotated on the basis of head-tracking data such that the head effectively remained fixed in space while the body rotated beneath it. Subjects updated most accurately in the passive condition but erred in the active and cancellation conditions. Performance is interpreted as reflecting the accuracy of perceived head rotation across conditions, which is modeled as a linear combination of proprioceptive/efference copy signals and vestibular signals. Resulting weights suggest that auditory updating is dominated by vestibular signals but with significant contributions from proprioception/efference copy. Overall, results shed light on the interplay of sensory and motor signals that determine the accuracy of auditory spatial updating. PMID:27169504

  19. Dependence of auditory spatial updating on vestibular, proprioceptive, and efference copy signals.

    PubMed

    Genzel, Daria; Firzlaff, Uwe; Wiegrebe, Lutz; MacNeilage, Paul R

    2016-08-01

    Humans localize sounds by comparing inputs across the two ears, resulting in a head-centered representation of sound-source position. When the head moves, information about head movement must be combined with the head-centered estimate to correctly update the world-centered sound-source position. Spatial updating has been extensively studied in the visual system, but less is known about how head movement signals interact with binaural information during auditory spatial updating. In the current experiments, listeners compared the world-centered azimuthal position of two sound sources presented before and after a head rotation that depended on condition. In the active condition, subjects rotated their head by ∼35° to the left or right, following a pretrained trajectory. In the passive condition, subjects were rotated along the same trajectory in a rotating chair. In the cancellation condition, subjects rotated their head as in the active condition, but the chair was counter-rotated on the basis of head-tracking data such that the head effectively remained fixed in space while the body rotated beneath it. Subjects updated most accurately in the passive condition but erred in the active and cancellation conditions. Performance is interpreted as reflecting the accuracy of perceived head rotation across conditions, which is modeled as a linear combination of proprioceptive/efference copy signals and vestibular signals. Resulting weights suggest that auditory updating is dominated by vestibular signals but with significant contributions from proprioception/efference copy. Overall, results shed light on the interplay of sensory and motor signals that determine the accuracy of auditory spatial updating. Copyright © 2016 the American Physiological Society.

  20. Investigation of the effects of a moving acoustic medium on jet noise measurements

    NASA Technical Reports Server (NTRS)

    Cole, J. E., III; Palmer, D. W.

    1976-01-01

    Noise from an unheated sonic jet in the presence of an external flow is measured in a free-jet wind tunnel using microphones located both inside and outside the flow. Comparison of the data is made with results of similar studies. The results are also compared with theoretical predictions of the source strength for jet noise in the presence of flow and of the effects of sound propagation through a shear layer.

  1. Intersensory Redundancy Facilitates Learning of Arbitrary Relations between Vowel Sounds and Objects in Seven-Month-Old Infants.

    ERIC Educational Resources Information Center

    Gogate, Lakshmi J.; Bahrick, Lorraine E.

    1998-01-01

    Investigated 7-month olds' ability to relate vowel sounds with objects when intersensory redundancy was present versus absent. Found that infants detected a mismatch in the vowel-object pairs in the moving-synchronous condition but not in the still or moving-asynchronous condition, demonstrating that temporal synchrony between vocalizations and…

  2. 76 FR 40617 - Security Zone; 2011 Seattle Seafair Fleet Week Moving Vessels, Puget Sound, Washington

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-07-11

    ...-AA87 Security Zone; 2011 Seattle Seafair Fleet Week Moving Vessels, Puget Sound, Washington AGENCY... security zones around the HMCS WHITEHORSE (NCSM 705), HMCS NANAIMO (NCSM 702), and the USCGC MELLON (WHEC... August 8, 2011. These security zones are necessary to help ensure the security of the vessels from...

  3. Effect of parameters in moving average method for event detection enhancement using phase sensitive OTDR

    NASA Astrophysics Data System (ADS)

    Kwon, Yong-Seok; Naeem, Khurram; Jeon, Min Yong; Kwon, Il-bum

    2017-04-01

    We analyze the relations of parameters in moving average method to enhance the event detectability of phase sensitive optical time domain reflectometer (OTDR). If the external events have unique frequency of vibration, then the control parameters of moving average method should be optimized in order to detect these events efficiently. A phase sensitive OTDR was implemented by a pulsed light source, which is composed of a laser diode, a semiconductor optical amplifier, an erbium-doped fiber amplifier, a fiber Bragg grating filter, and a light receiving part, which has a photo-detector and high speed data acquisition system. The moving average method is operated with the control parameters: total number of raw traces, M, number of averaged traces, N, and step size of moving, n. The raw traces are obtained by the phase sensitive OTDR with sound signals generated by a speaker. Using these trace data, the relation of the control parameters is analyzed. In the result, if the event signal has one frequency, then the optimal values of N, n are existed to detect the event efficiently.

  4. Psychophysical Evaluation of Three-Dimensional Auditory Displays

    NASA Technical Reports Server (NTRS)

    Wightman, Frederic L.

    1996-01-01

    This report describes the progress made during the second year of a three-year Cooperative Research Agreement. The CRA proposed a program of applied psychophysical research designed to determine the requirements and limitations of three-dimensional (3-D) auditory display systems. These displays present synthesized stimuli to a pilot or virtual workstation operator that evoke auditory images at predetermined positions in space. The images can be either stationary or moving. In previous years, we completed a number of studies that provided data on listeners' abilities to localize stationary sound sources with 3-D displays. The current focus is on the use of 3-D displays in 'natural' listening conditions, which include listeners'head movements, moving sources, multiple sources and 'echoic' sources. The results of our research on one of these topics, the localization of multiple sources, was reported in the most recent Semi-Annual Progress Report (Appendix A). That same progress report described work on two related topics, the influence of a listener's a-priori knowledge of source characteristics and the discriminability of real and virtual sources. In the period since the last Progress Report we have conducted several new studies to evaluate the effectiveness of a new and simpler method for measuring the HRTF's that are used to synthesize virtual sources and have expanded our studies of multiple sources. The results of this research are described below.

  5. On The Determination of Shadow Boundaries and Relevant Eigenrays For Sound Propagation In Stratified Moving Media

    NASA Technical Reports Server (NTRS)

    Mcaninch, Gerry L.; Heath, Stephanie L.

    2006-01-01

    The eiconal and ray equations for a stratified moving medium are analyzed to determine which rays connect a source located at an altitude z = z(sub s) to an observer on the ground plane. The characteristics of all available rays are ascertained. Further, the rays which determine the shadow boundary are found. Thus, the shadow boundary may be easily calculated and any observers for which an eigenray does not exist eliminated from further consideration. Therefore, the analysis may be applied to save computer time in two ways. First, rays that do not reach the ground may be eliminated before they are traced, and observers for which no eigenray exists may be eliminated from consideration.

  6. Numerical method to compute acoustic scattering effect of a moving source.

    PubMed

    Song, Hao; Yi, Mingxu; Huang, Jun; Pan, Yalin; Liu, Dawei

    2016-01-01

    In this paper, the aerodynamic characteristic of a ducted tail rotor in hover has been numerically studied using CFD method. An analytical time domain formulation based on Ffowcs Williams-Hawkings (FW-H) equation is derived for the prediction of the acoustic velocity field and used as Neumann boundary condition on a rigid scattering surface. In order to predict the aerodynamic noise, a hybrid method combing computational aeroacoustics with an acoustic thin-body boundary element method has been proposed. The aerodynamic results and the calculated sound pressure levels (SPLs) are compared with the known method for validation. Simulation results show that the duct can change the value of SPLs and the sound directivity. Compared with the isolate tail rotor, the SPLs of the ducted tail rotor are smaller at certain azimuth.

  7. Speech-on-speech masking in a front-back dimension and analysis of binaural parameters in rooms using MLS methods

    NASA Astrophysics Data System (ADS)

    Aaronson, Neil L.

    This dissertation deals with questions important to the problem of human sound source localization in rooms, starting with perceptual studies and moving on to physical measurements made in rooms. In Chapter 1, a perceptual study is performed relevant to a specific phenomenon the effect of speech reflections occurring in the front-back dimension and the ability of humans to segregate that from unreflected speech. Distracters were presented from the same source as the target speech, a loudspeaker directly in front of the listener, and also from a loudspeaker directly behind the listener, delayed relative to the front loudspeaker. Steps were taken to minimize the contributions of binaural difference cues. For all delays within +/-32 ms, a release from informational masking of about 2 dB occurred. This suggested that human listeners are able to segregate speech sources based on spatial cues, even with minimal binaural cues. In moving on to physical measurements in rooms, a method was sought for simultaneous measurement of room characteristics such as impulse response (IR) and reverberation time (RT60), and binaural parameters such as interaural time difference (ITD), interaural level difference (ILD), and the interaural cross-correlation function and coherence. Chapter 2 involves investigations into the usefulness of maximum length sequences (MLS) for these purposes. Comparisons to random telegraph noise (RTN) show that MLS performs better in the measurement of stationary and room transfer functions, IR, and RT60 by an order of magnitude in RMS percent error, even after Wiener filtering and exponential time-domain filtering have improved the accuracy of RTN measurements. Measurements were taken in real rooms in an effort to understand how the reverberant characteristics of rooms affect binaural parameters important to sound source localization. Chapter 3 deals with interaural coherence, a parameter important for localization and perception of auditory source width. MLS were used to measure waveform and envelope coherences in two rooms for various source distances and 0° azimuth through a head-and-torso simulator (KEMAR). A relationship is sought that relates these two types of coherence, since envelope coherence, while an important quantity, is generally less accessible than waveform coherence. A power law relationship is shown to exist between the two that works well within and across bands, for any source distance, and is robust to reverberant conditions of the room. Measurements of ITD, ILD, and coherence in rooms give insight into the way rooms affect these parameters, and in turn, the ability of listeners to localize sounds in rooms. Such measurements, along with room properties, are made and analyzed using MLS methods in Chapter 4. It was found that the pinnae cause incoherence for sound sources incident between 30° and 90°. In human listeners, this does not seem to adversely affect performance in lateralization experiments. The cause of poor coherence in rooms was studied as part of Chapter 4 as well. It was found that rooms affect coherence by introducing variance into the ITD spectra within the bands in which it is measured. A mathematical model to predict the interaural coherence within a band given the standard deviation of the ITD spectrum and the center frequency of the band gives an exponential relationship. This is found to work well in predicting measured coherence given ITD spectrum variance. The pinnae seem to affect the ITD spectrum in a similar way at incident sound angles for which coherence is poor in an anechoic environment.

  8. Just How Does Sound Wave?

    ERIC Educational Resources Information Center

    Shipman, Bob

    2006-01-01

    When children first hear the term "sound wave" perhaps they might associate it with the way a hand waves or perhaps the squiggly line image on a television monitor when sound recordings are being made. Research suggests that children tend to think sound somehow travels as a discrete package, a fast-moving invisible thing, and not something that…

  9. Self-generated sounds of locomotion and ventilation and the evolution of human rhythmic abilities.

    PubMed

    Larsson, Matz

    2014-01-01

    It has been suggested that the basic building blocks of music mimic sounds of moving humans, and because the brain was primed to exploit such sounds, they eventually became incorporated in human culture. However, that raises further questions. Why do genetically close, culturally well-developed apes lack musical abilities? Did our switch to bipedalism influence the origins of music? Four hypotheses are raised: (1) Human locomotion and ventilation can mask critical sounds in the environment. (2) Synchronization of locomotion reduces that problem. (3) Predictable sounds of locomotion may stimulate the evolution of synchronized behavior. (4) Bipedal gait and the associated sounds of locomotion influenced the evolution of human rhythmic abilities. Theoretical models and research data suggest that noise of locomotion and ventilation may mask critical auditory information. People often synchronize steps subconsciously. Human locomotion is likely to produce more predictable sounds than those of non-human primates. Predictable locomotion sounds may have improved our capacity of entrainment to external rhythms and to feel the beat in music. A sense of rhythm could aid the brain in distinguishing among sounds arising from discrete sources and also help individuals to synchronize their movements with one another. Synchronization of group movement may improve perception by providing periods of relative silence and by facilitating auditory processing. The adaptive value of such skills to early ancestors may have been keener detection of prey or stalkers and enhanced communication. Bipedal walking may have influenced the development of entrainment in humans and thereby the evolution of rhythmic abilities.

  10. Judging sound rotation when listeners and sounds rotate: Sound source localization is a multisystem process.

    PubMed

    Yost, William A; Zhong, Xuan; Najam, Anbar

    2015-11-01

    In four experiments listeners were rotated or were stationary. Sounds came from a stationary loudspeaker or rotated from loudspeaker to loudspeaker around an azimuth array. When either sounds or listeners rotate the auditory cues used for sound source localization change, but in the everyday world listeners perceive sound rotation only when sounds rotate not when listeners rotate. In the everyday world sound source locations are referenced to positions in the environment (a world-centric reference system). The auditory cues for sound source location indicate locations relative to the head (a head-centric reference system), not locations relative to the world. This paper deals with a general hypothesis that the world-centric location of sound sources requires the auditory system to have information about auditory cues used for sound source location and cues about head position. The use of visual and vestibular information in determining rotating head position in sound rotation perception was investigated. The experiments show that sound rotation perception when sources and listeners rotate was based on acoustic, visual, and, perhaps, vestibular information. The findings are consistent with the general hypotheses and suggest that sound source localization is not based just on acoustics. It is a multisystem process.

  11. Mathematically trivial control of sound using a parametric beam focusing source.

    PubMed

    Tanaka, Nobuo; Tanaka, Motoki

    2011-01-01

    By exploiting a case regarded as trivial, this paper presents global active noise control using a parametric beam focusing source (PBFS). As with a dipole model, one is used for a primary sound source and the other for a control sound source, the control effect for minimizing a total acoustic power depends on the distance between the two. When the distance becomes zero, the total acoustic power becomes null, hence nothing less than a trivial case. Because of the constraints in practice, there exist difficulties in placing a control source close enough to a primary source. However, by projecting a sound beam of a parametric array loudspeaker onto the target sound source (primary source), a virtual sound source may be created on the target sound source, thereby enabling the collocation of the sources. In order to further ensure feasibility of the trivial case, a PBFS is then introduced in an effort to meet the size of the two sources. Reflected sound wave of the PBFS, which is tantamount to the virtual sound source output, aims to suppress the primary sound. Finally, a numerical analysis as well as an experiment is conducted, verifying the validity of the proposed methodology.

  12. A Flexible 360-Degree Thermal Sound Source Based on Laser Induced Graphene

    PubMed Central

    Tao, Lu-Qi; Liu, Ying; Ju, Zhen-Yi; Tian, He; Xie, Qian-Yi; Yang, Yi; Ren, Tian-Ling

    2016-01-01

    A flexible sound source is essential in a whole flexible system. It’s hard to integrate a conventional sound source based on a piezoelectric part into a whole flexible system. Moreover, the sound pressure from the back side of a sound source is usually weaker than that from the front side. With the help of direct laser writing (DLW) technology, the fabrication of a flexible 360-degree thermal sound source becomes possible. A 650-nm low-power laser was used to reduce the graphene oxide (GO). The stripped laser induced graphene thermal sound source was then attached to the surface of a cylindrical bottle so that it could emit sound in a 360-degree direction. The sound pressure level and directivity of the sound source were tested, and the results were in good agreement with the theoretical results. Because of its 360-degree sound field, high flexibility, high efficiency, low cost, and good reliability, the 360-degree thermal acoustic sound source will be widely applied in consumer electronics, multi-media systems, and ultrasonic detection and imaging. PMID:28335239

  13. Exact solutions for sound radiation from a moving monopole above an impedance plane.

    PubMed

    Ochmann, Martin

    2013-04-01

    The acoustic field of a monopole source moving with constant velocity at constant height above an infinite locally reacting plane can be expressed in analytical form by combining the Lorentz transformation with the method of superimposing complex or real point sources. For a plane with masslike response, the solution in Lorentz space consists of a superposition of monopoles only and therefore, does not differ in principle from the solution for the corresponding stationary boundary value problem. However, by considering a frequency independent surface impedance, e.g., with pure absorbing behavior, the half-space Green's function is now comprised of not only a line of monopoles but also of dipoles. For certain field points at a special line g, this solution can be written explicitly by using an exponential integral. For arbitrary field points, the method of stationary phase leads to an asymptotic solution for the reflection coefficient which agrees with prior results from the literature.

  14. Direct-location versus verbal report methods for measuring auditory distance perception in the far field.

    PubMed

    Etchemendy, Pablo E; Spiousas, Ignacio; Calcagno, Esteban R; Abregú, Ezequiel; Eguia, Manuel C; Vergara, Ramiro O

    2018-06-01

    In this study we evaluated whether a method of direct location is an appropriate response method for measuring auditory distance perception of far-field sound sources. We designed an experimental set-up that allows participants to indicate the distance at which they perceive the sound source by moving a visual marker. We termed this method Cross-Modal Direct Location (CMDL) since the response procedure involves the visual modality while the stimulus is presented through the auditory modality. Three experiments were conducted with sound sources located from 1 to 6 m. The first one compared the perceived distances obtained using either the CMDL device or verbal report (VR), which is the response method more frequently used for reporting auditory distance in the far field, and found differences on response compression and bias. In Experiment 2, participants reported visual distance estimates to the visual marker that were found highly accurate. Then, we asked the same group of participants to report VR estimates of auditory distance and found that the spatial visual information, obtained from the previous task, did not influence their reports. Finally, Experiment 3 compared the same responses that Experiment 1 but interleaving the methods, showing a weak, but complex, mutual influence. However, the estimates obtained with each method remained statistically different. Our results show that the auditory distance psychophysical functions obtained with the CMDL method are less susceptible to previously reported underestimation for distances over 2 m.

  15. The Relative Contribution of Interaural Time and Magnitude Cues to Dynamic Sound Localization

    NASA Technical Reports Server (NTRS)

    Wenzel, Elizabeth M.; Null, Cynthia H. (Technical Monitor)

    1995-01-01

    This paper presents preliminary data from a study examining the relative contribution of interaural time differences (ITDs) and interaural level differences (ILDs) to the localization of virtual sound sources both with and without head motion. The listeners' task was to estimate the apparent direction and distance of virtual sources (broadband noise) presented over headphones. Stimuli were synthesized from minimum phase representations of nonindividualized directional transfer functions; binaural magnitude spectra were derived from the minimum phase estimates and ITDs were represented as a pure delay. During dynamic conditions, listeners were encouraged to move their heads; the position of the listener's head was tracked and the stimuli were synthesized in real time using a Convolvotron to simulate a stationary external sound source. ILDs and ITDs were either correctly or incorrectly correlated with head motion: (1) both ILDs and ITDs correctly correlated, (2) ILDs correct, ITD fixed at 0 deg azimuth and 0 deg elevation, (3) ITDs correct, ILDs fixed at 0 deg, 0 deg. Similar conditions were run for static conditions except that none of the cues changed with head motion. The data indicated that, compared to static conditions, head movements helped listeners to resolve confusions primarily when ILDs were correctly correlated, although a smaller effect was also seen for correct ITDs. Together with the results for static conditions, the data suggest that localization tends to be dominated by the cue that is most reliable or consistent, when reliability is defined by consistency over time as well as across frequency bands.

  16. Trading of dynamic interaural time and level difference cues and its effect on the auditory motion-onset response measured with electroencephalography.

    PubMed

    Altmann, Christian F; Ueda, Ryuhei; Bucher, Benoit; Furukawa, Shigeto; Ono, Kentaro; Kashino, Makio; Mima, Tatsuya; Fukuyama, Hidenao

    2017-10-01

    Interaural time (ITD) and level differences (ILD) constitute the two main cues for sound localization in the horizontal plane. Despite extensive research in animal models and humans, the mechanism of how these two cues are integrated into a unified percept is still far from clear. In this study, our aim was to test with human electroencephalography (EEG) whether integration of dynamic ITD and ILD cues is reflected in the so-called motion-onset response (MOR), an evoked potential elicited by moving sound sources. To this end, ITD and ILD trajectories were determined individually by cue trading psychophysics. We then measured EEG while subjects were presented with either static click-trains or click-trains that contained a dynamic portion at the end. The dynamic part was created by combining ITD with ILD either congruently to elicit the percept of a right/leftward moving sound, or incongruently to elicit the percept of a static sound. In two experiments that differed in the method to derive individual dynamic cue trading stimuli, we observed an MOR with at least a change-N1 (cN1) component for both the congruent and incongruent conditions at about 160-190 ms after motion-onset. A significant change-P2 (cP2) component for both the congruent and incongruent ITD/ILD combination was found only in the second experiment peaking at about 250 ms after motion onset. In sum, this study shows that a sound which - by a combination of counter-balanced ITD and ILD cues - induces a static percept can still elicit a motion-onset response, indicative of independent ITD and ILD processing at the level of the MOR - a component that has been proposed to be, at least partly, generated in non-primary auditory cortex. Copyright © 2017 Elsevier Inc. All rights reserved.

  17. Acoustic radiosity for computation of sound fields in diffuse environments

    NASA Astrophysics Data System (ADS)

    Muehleisen, Ralph T.; Beamer, C. Walter

    2002-05-01

    The use of image and ray tracing methods (and variations thereof) for the computation of sound fields in rooms is relatively well developed. In their regime of validity, both methods work well for prediction in rooms with small amounts of diffraction and mostly specular reflection at the walls. While extensions to the method to include diffuse reflections and diffraction have been made, they are limited at best. In the fields of illumination and computer graphics the ray tracing and image methods are joined by another method called luminous radiative transfer or radiosity. In radiosity, an energy balance between surfaces is computed assuming diffuse reflection at the reflective surfaces. Because the interaction between surfaces is constant, much of the computation required for sound field prediction with multiple or moving source and receiver positions can be reduced. In acoustics the radiosity method has had little attention because of the problems of diffraction and specular reflection. The utility of radiosity in acoustics and an approach to a useful development of the method for acoustics will be presented. The method looks especially useful for sound level prediction in industrial and office environments. [Work supported by NSF.

  18. Psychophysical Evaluation of Three-Dimensional Auditory Displays

    NASA Technical Reports Server (NTRS)

    Wightman, Frederic L. (Principal Investigator)

    1995-01-01

    This report describes the process made during the first year of a three-year Cooperative Research Agreement (CRA NCC2-542). The CRA proposed a program of applied of psychophysical research designed to determine the requirements and limitations of three-dimensional (3-D) auditory display systems. These displays present synthesized stimuli to a pilot or virtual workstation operator that evoke auditory images at predetermined positions in space. The images can be either stationary or moving. In previous years. we completed a number of studies that provided data on listeners' abilities to localize stationary sound sources with 3-D displays. The current focus is on the use of 3-D displays in 'natural' listening conditions, which include listeners' head movements, moving sources, multiple sources and 'echoic' sources. The results of our research on two of these topics, the role of head movements and the role of echoes and reflections, were reported in the most recent Semi-Annual Pro-ress Report (Appendix A). In the period since the last Progress Report we have been studying a third topic, the localizability of moving sources. The results of this research are described. The fidelity of a virtual auditory display is critically dependent on precise measurement of the listener''s Head-Related Transfer Functions (HRTFs), which are used to produce the virtual auditory images. We continue to explore methods for improving our HRTF measurement technique. During this reporting period we compared HRTFs measured using our standard open-canal probe tube technique and HRTFs measured with the closed-canal insert microphones from the Crystal River Engineering Snapshot system.

  19. Characterizing, synthesizing, and/or canceling out acoustic signals from sound sources

    DOEpatents

    Holzrichter, John F [Berkeley, CA; Ng, Lawrence C [Danville, CA

    2007-03-13

    A system for characterizing, synthesizing, and/or canceling out acoustic signals from inanimate and animate sound sources. Electromagnetic sensors monitor excitation sources in sound producing systems, such as animate sound sources such as the human voice, or from machines, musical instruments, and various other structures. Acoustical output from these sound producing systems is also monitored. From such information, a transfer function characterizing the sound producing system is generated. From the transfer function, acoustical output from the sound producing system may be synthesized or canceled. The systems disclosed enable accurate calculation of transfer functions relating specific excitations to specific acoustical outputs. Knowledge of such signals and functions can be used to effect various sound replication, sound source identification, and sound cancellation applications.

  20. Moving microphone arrays to reduce spatial aliasing in the beamforming technique: theoretical background and numerical investigation.

    PubMed

    Cigada, Alfredo; Lurati, Massimiliano; Ripamonti, Francesco; Vanali, Marcello

    2008-12-01

    This paper introduces a measurement technique aimed at reducing or possibly eliminating the spatial aliasing problem in the beamforming technique. Beamforming main disadvantages are a poor spatial resolution, at low frequency, and the spatial aliasing problem, at higher frequency, leading to the identification of false sources. The idea is to move the microphone array during the measurement operation. In this paper, the proposed approach is theoretically and numerically investigated by means of simple sound propagation models, proving its efficiency in reducing the spatial aliasing. A number of different array configurations are numerically investigated together with the most important parameters governing this measurement technique. A set of numerical results concerning the case of a planar rotating array is shown, together with a first experimental validation of the method.

  1. The aeromagnetic method as a tool to identify Cenozoic magmatism in the West Antarctic Rift System beneath the West Antarctic Ice Sheet: a review; Thiel subglacial volcano as possible source of the ash layer in the WAISCOR

    USGS Publications Warehouse

    Behrendt, John C.

    2013-01-01

    The West Antarctic Ice Sheet (WAIS) flows through the volcanically active West Antarctic Rift System (WARS). The aeromagnetic method has been the most useful geophysical tool for identification of subglacial volcanic rocks, since 1959–64 surveys, particularly combined with 1978 radar ice-sounding. The unique 1991–97 Central West Antarctica (CWA) aerogeophysical survey covering 354,000 km2 over the WAIS, (5-km line-spaced, orthogonal lines of aeromagnetic, radar ice-sounding, and aerogravity measurements), still provides invaluable information on subglacial volcanic rocks, particularly combined with the older aeromagnetic profiles. These data indicate numerous 100–>1000 nT, 5–50-km width, shallow-source, magnetic anomalies over an area greater than 1.2 × 106 km2, mostly from subglacial volcanic sources. I interpreted the CWA anomalies as defining about 1000 “volcanic centers” requiring high remanent normal magnetizations in the present field direction. About 400 anomaly sources correlate with bed topography. At least 80% of these sources have less than 200 m relief at the WAIS bed. They appear modified by moving ice, requiring a younger age than the WAIS (about 25 Ma). Exposed volcanoes in the WARS are The present rapid changes resulting from global warming, could be accelerated by subglacial volcanism.

  2. A numerical study of fundamental shock noise mechanisms. Ph.D. Thesis - Cornell Univ.

    NASA Technical Reports Server (NTRS)

    Meadows, Kristine R.

    1995-01-01

    The results of this thesis demonstrate that direct numerical simulation can predict sound generation in unsteady aerodynamic flows containing shock waves. Shock waves can be significant sources of sound in high speed jet flows, on helicopter blades, and in supersonic combustion inlets. Direct computation of sound permits the prediction of noise levels in the preliminary design stage and can be used as a tool to focus experimental studies, thereby reducing cost and increasing the probability of a successfully quiet product in less time. This thesis reveals and investigates two mechanisms fundamental to sound generation by shocked flows: shock motion and shock deformation. Shock motion is modeled by the interaction of a sound wave with a shock. During the interaction, the shock wave begins to move and the sound pressure is amplified as the wave passes through the shock. The numerical approach presented in this thesis is validated by the comparison of results obtained in a quasi-one dimensional simulation with linear theory. Analysis of the perturbation energy demonstrated for the first time that acoustic energy is generated by the interaction. Shock deformation is investigated by the numerical simulation of a ring vortex interacting with a shock. This interaction models the passage of turbulent structures through the shock wave. The simulation demonstrates that both acoustic waves and contact surfaces are generated downstream during the interaction. Analysis demonstrates that the acoustic wave spreads cylindrically, that the sound intensity is highly directional, and that the sound pressure level increases significantly with increasing shock strength. The effect of shock strength on sound pressure level is consistent with experimental observations of shock noise, indicating that the interaction of a ring vortex with a shock wave correctly models a dominant mechanism of shock noise generation.

  3. Directional Hearing and Sound Source Localization in Fishes.

    PubMed

    Sisneros, Joseph A; Rogers, Peter H

    2016-01-01

    Evidence suggests that the capacity for sound source localization is common to mammals, birds, reptiles, and amphibians, but surprisingly it is not known whether fish locate sound sources in the same manner (e.g., combining binaural and monaural cues) or what computational strategies they use for successful source localization. Directional hearing and sound source localization in fishes continues to be important topics in neuroethology and in the hearing sciences, but the empirical and theoretical work on these topics have been contradictory and obscure for decades. This chapter reviews the previous behavioral work on directional hearing and sound source localization in fishes including the most recent experiments on sound source localization by the plainfin midshipman fish (Porichthys notatus), which has proven to be an exceptional species for fish studies of sound localization. In addition, the theoretical models of directional hearing and sound source localization for fishes are reviewed including a new model that uses a time-averaged intensity approach for source localization that has wide applicability with regard to source type, acoustic environment, and time waveform.

  4. Sounds scary? Lack of habituation following the presentation of novel sounds.

    PubMed

    Biedenweg, Tine A; Parsons, Michael H; Fleming, Patricia A; Blumstein, Daniel T

    2011-01-18

    Animals typically show less habituation to biologically meaningful sounds than to novel signals. We might therefore expect that acoustic deterrents should be based on natural sounds. We investigated responses by western grey kangaroos (Macropus fulignosus) towards playback of natural sounds (alarm foot stomps and Australian raven (Corvus coronoides) calls) and artificial sounds (faux snake hiss and bull whip crack). We then increased rate of presentation to examine whether animals would habituate. Finally, we varied frequency of playback to investigate optimal rates of delivery. Nine behaviors clustered into five Principal Components. PC factors 1 and 2 (animals alert or looking, or hopping and moving out of area) accounted for 36% of variance. PC factor 3 (eating cessation, taking flight, movement out of area) accounted for 13% of variance. Factors 4 and 5 (relaxing, grooming and walking; 12 and 11% of variation, respectively) discontinued upon playback. The whip crack was most evocative; eating was reduced from 75% of time spent prior to playback to 6% following playback (post alarm stomp: 32%, raven call: 49%, hiss: 75%). Additionally, 24% of individuals took flight and moved out of area (50 m radius) in response to the whip crack (foot stomp: 0%, raven call: 8% and 4%, hiss: 6%). Increasing rate of presentation (12x/min ×2 min) caused 71% of animals to move out of the area. The bull whip crack, an artificial sound, was as effective as the alarm stomp at eliciting aversive behaviors. Kangaroos did not fully habituate despite hearing the signal up to 20x/min. Highest rates of playback did not elicit the greatest responses, suggesting that 'more is not always better'. Ultimately, by utilizing both artificial and biological sounds, predictability may be masked or offset, so that habituation is delayed and more effective deterrents may be produced.

  5. Environmental Assessment: Installation Development at Sheppard Air Force Base, Texas

    DTIC Science & Technology

    2007-05-01

    column, or in topographic depressions. Water is then utilized by plants and is respired, or it moves slowly into groundwater and/or eventually to surface...water bodies where it slowly moves through the hydrologic cycle. Removal of vegetation decreases infiltration into the soil column and thereby...School District JP-4 jet propulsion fuel 4 kts knots Ldn Day- Night Average Sound Level Leq equivalent noise level Lmax maximum sound level lb pound

  6. Source and listener directivity for interactive wave-based sound propagation.

    PubMed

    Mehra, Ravish; Antani, Lakulish; Kim, Sujeong; Manocha, Dinesh

    2014-04-01

    We present an approach to model dynamic, data-driven source and listener directivity for interactive wave-based sound propagation in virtual environments and computer games. Our directional source representation is expressed as a linear combination of elementary spherical harmonic (SH) sources. In the preprocessing stage, we precompute and encode the propagated sound fields due to each SH source. At runtime, we perform the SH decomposition of the varying source directivity interactively and compute the total sound field at the listener position as a weighted sum of precomputed SH sound fields. We propose a novel plane-wave decomposition approach based on higher-order derivatives of the sound field that enables dynamic HRTF-based listener directivity at runtime. We provide a generic framework to incorporate our source and listener directivity in any offline or online frequency-domain wave-based sound propagation algorithm. We have integrated our sound propagation system in Valve's Source game engine and use it to demonstrate realistic acoustic effects such as sound amplification, diffraction low-passing, scattering, localization, externalization, and spatial sound, generated by wave-based propagation of directional sources and listener in complex scenarios. We also present results from our preliminary user study.

  7. Your Child's Development: 1 Month

    MedlinePlus

    ... becomes alert when hearing a pleasant sound, like music Cognitive Skills (Thinking and Learning) will stare at ... pleasing sounds, such as "baby talk" or gentle music has legs or hands that do not move ...

  8. Response of humpback whales (Megaptera novaeangliae) to ramp-up of a small experimental air gun array.

    PubMed

    Dunlop, Rebecca A; Noad, Michael J; McCauley, Robert D; Kniest, Eric; Slade, Robert; Paton, David; Cato, Douglas H

    2016-02-15

    'Ramp-up', or 'soft start', is a mitigation measure used in seismic surveys and involves increasing the radiated sound level over 20-40 min. This study compared the behavioural response in migrating humpback whales to the first stages of ramp-up with the response to a 'constant' source, 'controls' (in which the array was towed but not operated) with groups in the absence of the source vessel used as the 'baseline'. Although the behavioural response, in most groups, resulted in an increase in distance from the source (potential avoidance), there was no evidence that either 'ramp-up' or the constant source at a higher level was superior for triggering whales to move away from the source vessel. 'Control' groups also responded suggesting the presence of the source vessel had some effect. However, the majority of groups appeared to avoid the source vessel at distances greater than the radius of most mitigation zones. Copyright © 2016 Elsevier Ltd. All rights reserved.

  9. Virtual environment display for a 3D audio room simulation

    NASA Astrophysics Data System (ADS)

    Chapin, William L.; Foster, Scott

    1992-06-01

    Recent developments in virtual 3D audio and synthetic aural environments have produced a complex acoustical room simulation. The acoustical simulation models a room with walls, ceiling, and floor of selected sound reflecting/absorbing characteristics and unlimited independent localizable sound sources. This non-visual acoustic simulation, implemented with 4 audio ConvolvotronsTM by Crystal River Engineering and coupled to the listener with a Poihemus IsotrakTM, tracking the listener's head position and orientation, and stereo headphones returning binaural sound, is quite compelling to most listeners with eyes closed. This immersive effect should be reinforced when properly integrated into a full, multi-sensory virtual environment presentation. This paper discusses the design of an interactive, visual virtual environment, complementing the acoustic model and specified to: 1) allow the listener to freely move about the space, a room of manipulable size, shape, and audio character, while interactively relocating the sound sources; 2) reinforce the listener's feeling of telepresence into the acoustical environment with visual and proprioceptive sensations; 3) enhance the audio with the graphic and interactive components, rather than overwhelm or reduce it; and 4) serve as a research testbed and technology transfer demonstration. The hardware/software design of two demonstration systems, one installed and one portable, are discussed through the development of four iterative configurations. The installed system implements a head-coupled, wide-angle, stereo-optic tracker/viewer and multi-computer simulation control. The portable demonstration system implements a head-mounted wide-angle, stereo-optic display, separate head and pointer electro-magnetic position trackers, a heterogeneous parallel graphics processing system, and object oriented C++ program code.

  10. Transcutaneous measurement of volume blood flow

    NASA Technical Reports Server (NTRS)

    Daigle, R. E.; Mcleod, F. D.; Miller, C. W.; Histand, M. B.; Wells, M. K.

    1974-01-01

    Blood flow velocity measurements, using Doppler velocimeter, are described. The ability to measure blood velocity using ultrasound is derived from the Doppler effect; the change in frequency which occurs when sound is reflected or transmitted from a moving target. When ultrasound of the appropriate frequency is transmitted through a moving blood stream, the blood cells act as point scatterers of ultrasonic energy. If this scattered ultrasonic energy is detected, it is found to be shifted in frequency according to the velocity of the blood cells, nu, the frequency of the incident sound, f sub o, the speed of sound in the medium, c, and the angle between the sound beam and the velocity vector, o. The relation describing this effect is known as the Doppler equation. Delta f = 2 f sub o x nu x cos alpha/c. The theoretical and experimental methods are evaluated.

  11. The moving minimum audible angle is smaller during self motion than during source motion

    PubMed Central

    Brimijoin, W. Owen; Akeroyd, Michael A.

    2014-01-01

    We are rarely perfectly still: our heads rotate in three axes and move in three dimensions, constantly varying the spectral and binaural cues at the ear drums. In spite of this motion, static sound sources in the world are typically perceived as stable objects. This argues that the auditory system—in a manner not unlike the vestibulo-ocular reflex—works to compensate for self motion and stabilize our sensory representation of the world. We tested a prediction arising from this postulate: that self motion should be processed more accurately than source motion. We used an infrared motion tracking system to measure head angle, and real-time interpolation of head related impulse responses to create “head-stabilized” signals that appeared to remain fixed in space as the head turned. After being presented with pairs of simultaneous signals consisting of a man and a woman speaking a snippet of speech, normal and hearing impaired listeners were asked to report whether the female voice was to the left or the right of the male voice. In this way we measured the moving minimum audible angle (MMAA). This measurement was made while listeners were asked to turn their heads back and forth between ± 15° and the signals were stabilized in space. After this “self-motion” condition we measured MMAA in a second “source-motion” condition when listeners remained still and the virtual locations of the signals were moved using the trajectories from the first condition. For both normal and hearing impaired listeners, we found that the MMAA for signals moving relative to the head was ~1–2° smaller when the movement was the result of self motion than when it was the result of source motion, even though the motion with respect to the head was identical. These results as well as the results of past experiments suggest that spatial processing involves an ongoing and highly accurate comparison of spatial acoustic cues with self-motion cues. PMID:25228856

  12. Selective Listening Point Audio Based on Blind Signal Separation and Stereophonic Technology

    NASA Astrophysics Data System (ADS)

    Niwa, Kenta; Nishino, Takanori; Takeda, Kazuya

    A sound field reproduction method is proposed that uses blind source separation and a head-related transfer function. In the proposed system, multichannel acoustic signals captured at distant microphones are decomposed to a set of location/signal pairs of virtual sound sources based on frequency-domain independent component analysis. After estimating the locations and the signals of the virtual sources by convolving the controlled acoustic transfer functions with each signal, the spatial sound is constructed at the selected point. In experiments, a sound field made by six sound sources is captured using 48 distant microphones and decomposed into sets of virtual sound sources. Since subjective evaluation shows no significant difference between natural and reconstructed sound when six virtual sources and are used, the effectiveness of the decomposing algorithm as well as the virtual source representation are confirmed.

  13. A method for evaluating the relation between sound source segregation and masking

    PubMed Central

    Lutfi, Robert A.; Liu, Ching-Ju

    2011-01-01

    Sound source segregation refers to the ability to hear as separate entities two or more sound sources comprising a mixture. Masking refers to the ability of one sound to make another sound difficult to hear. Often in studies, masking is assumed to result from a failure of segregation, but this assumption may not always be correct. Here a method is offered to identify the relation between masking and sound source segregation in studies and an example is given of its application. PMID:21302979

  14. Sound Sources Identified in High-Speed Jets by Correlating Flow Density Fluctuations With Far-Field Noise

    NASA Technical Reports Server (NTRS)

    Panda, Jayanta; Seasholtz, Richard G.

    2003-01-01

    Noise sources in high-speed jets were identified by directly correlating flow density fluctuation (cause) to far-field sound pressure fluctuation (effect). The experimental study was performed in a nozzle facility at the NASA Glenn Research Center in support of NASA s initiative to reduce the noise emitted by commercial airplanes. Previous efforts to use this correlation method have failed because the tools for measuring jet turbulence were intrusive. In the present experiment, a molecular Rayleigh-scattering technique was used that depended on laser light scattering by gas molecules in air. The technique allowed accurate measurement of air density fluctuations from different points in the plume. The study was conducted in shock-free, unheated jets of Mach numbers 0.95, 1.4, and 1.8. The turbulent motion, as evident from density fluctuation spectra was remarkably similar in all three jets, whereas the noise sources were significantly different. The correlation study was conducted by keeping a microphone at a fixed location (at the peak noise emission angle of 30 to the jet axis and 50 nozzle diameters away) while moving the laser probe volume from point to point in the flow. The following figure shows maps of the nondimensional coherence value measured at different Strouhal frequencies ([frequency diameter]/jet speed) in the supersonic Mach 1.8 and subsonic Mach 0.95 jets. The higher the coherence, the stronger the source was.

  15. Acoustic change detection algorithm using an FM radio

    NASA Astrophysics Data System (ADS)

    Goldman, Geoffrey H.; Wolfe, Owen

    2012-06-01

    The U.S. Army is interested in developing low-cost, low-power, non-line-of-sight sensors for monitoring human activity. One modality that is often overlooked is active acoustics using sources of opportunity such as speech or music. Active acoustics can be used to detect human activity by generating acoustic images of an area at different times, then testing for changes among the imagery. A change detection algorithm was developed to detect physical changes in a building, such as a door changing positions or a large box being moved using acoustics sources of opportunity. The algorithm is based on cross correlating the acoustic signal measured from two microphones. The performance of the algorithm was shown using data generated with a hand-held FM radio as a sound source and two microphones. The algorithm could detect a door being opened in a hallway.

  16. Sound source localization identification accuracy: Envelope dependencies.

    PubMed

    Yost, William A

    2017-07-01

    Sound source localization accuracy as measured in an identification procedure in a front azimuth sound field was studied for click trains, modulated noises, and a modulated tonal carrier. Sound source localization accuracy was determined as a function of the number of clicks in a 64 Hz click train and click rate for a 500 ms duration click train. The clicks were either broadband or high-pass filtered. Sound source localization accuracy was also measured for a single broadband filtered click and compared to a similar broadband filtered, short-duration noise. Sound source localization accuracy was determined as a function of sinusoidal amplitude modulation and the "transposed" process of modulation of filtered noises and a 4 kHz tone. Different rates (16 to 512 Hz) of modulation (including unmodulated conditions) were used. Providing modulation for filtered click stimuli, filtered noises, and the 4 kHz tone had, at most, a very small effect on sound source localization accuracy. These data suggest that amplitude modulation, while providing information about interaural time differences in headphone studies, does not have much influence on sound source localization accuracy in a sound field.

  17. System and method for characterizing synthesizing and/or canceling out acoustic signals from inanimate sound sources

    DOEpatents

    Holzrichter, John F.; Burnett, Greg C.; Ng, Lawrence C.

    2003-01-01

    A system and method for characterizing, synthesizing, and/or canceling out acoustic signals from inanimate sound sources is disclosed. Propagating wave electromagnetic sensors monitor excitation sources in sound producing systems, such as machines, musical instruments, and various other structures. Acoustical output from these sound producing systems is also monitored. From such information, a transfer function characterizing the sound producing system is generated. From the transfer function, acoustical output from the sound producing system may be synthesized or canceled. The methods disclosed enable accurate calculation of matched transfer functions relating specific excitations to specific acoustical outputs. Knowledge of such signals and functions can be used to effect various sound replication, sound source identification, and sound cancellation applications.

  18. System and method for characterizing, synthesizing, and/or canceling out acoustic signals from inanimate sound sources

    DOEpatents

    Holzrichter, John F; Burnett, Greg C; Ng, Lawrence C

    2013-05-21

    A system and method for characterizing, synthesizing, and/or canceling out acoustic signals from inanimate sound sources is disclosed. Propagating wave electromagnetic sensors monitor excitation sources in sound producing systems, such as machines, musical instruments, and various other structures. Acoustical output from these sound producing systems is also monitored. From such information, a transfer function characterizing the sound producing system is generated. From the transfer function, acoustical output from the sound producing system may be synthesized or canceled. The methods disclosed enable accurate calculation of matched transfer functions relating specific excitations to specific acoustical outputs. Knowledge of such signals and functions can be used to effect various sound replication, sound source identification, and sound cancellation applications.

  19. System and method for characterizing, synthesizing, and/or canceling out acoustic signals from inanimate sound sources

    DOEpatents

    Holzrichter, John F.; Burnett, Greg C.; Ng, Lawrence C.

    2007-10-16

    A system and method for characterizing, synthesizing, and/or canceling out acoustic signals from inanimate sound sources is disclosed. Propagating wave electromagnetic sensors monitor excitation sources in sound producing systems, such as machines, musical instruments, and various other structures. Acoustical output from these sound producing systems is also monitored. From such information, a transfer function characterizing the sound producing system is generated. From the transfer function, acoustical output from the sound producing system may be synthesized or canceled. The methods disclosed enable accurate calculation of matched transfer functions relating specific excitations to specific acoustical outputs. Knowledge of such signals and functions can be used to effect various sound replication, sound source identification, and sound cancellation applications.

  20. SoundCompass: A Distributed MEMS Microphone Array-Based Sensor for Sound Source Localization

    PubMed Central

    Tiete, Jelmer; Domínguez, Federico; da Silva, Bruno; Segers, Laurent; Steenhaut, Kris; Touhafi, Abdellah

    2014-01-01

    Sound source localization is a well-researched subject with applications ranging from localizing sniper fire in urban battlefields to cataloging wildlife in rural areas. One critical application is the localization of noise pollution sources in urban environments, due to an increasing body of evidence linking noise pollution to adverse effects on human health. Current noise mapping techniques often fail to accurately identify noise pollution sources, because they rely on the interpolation of a limited number of scattered sound sensors. Aiming to produce accurate noise pollution maps, we developed the SoundCompass, a low-cost sound sensor capable of measuring local noise levels and sound field directionality. Our first prototype is composed of a sensor array of 52 Microelectromechanical systems (MEMS) microphones, an inertial measuring unit and a low-power field-programmable gate array (FPGA). This article presents the SoundCompass’s hardware and firmware design together with a data fusion technique that exploits the sensing capabilities of the SoundCompass in a wireless sensor network to localize noise pollution sources. Live tests produced a sound source localization accuracy of a few centimeters in a 25-m2 anechoic chamber, while simulation results accurately located up to five broadband sound sources in a 10,000-m2 open field. PMID:24463431

  1. Pupils Think Sound Has Substance--Well, Sort of ...

    ERIC Educational Resources Information Center

    Whittaker, Andrew G.

    2012-01-01

    Physics is a subject where pupils hold a great number of deeply seated misconceptions. Sound is a prime example, as it requires the visualisation of a form of energy that moves imperceptibly through an invisible medium. This article outlines some of the common misconceptions that pupils hold regarding the nature of sound and how it is transmitted,…

  2. Spherical loudspeaker array for local active control of sound.

    PubMed

    Rafaely, Boaz

    2009-05-01

    Active control of sound has been employed to reduce noise levels around listeners' head using destructive interference from noise-canceling sound sources. Recently, spherical loudspeaker arrays have been studied as multiple-channel sound sources, capable of generating sound fields with high complexity. In this paper, the potential use of a spherical loudspeaker array for local active control of sound is investigated. A theoretical analysis of the primary and secondary sound fields around a spherical sound source reveals that the natural quiet zones for the spherical source have a shell-shape. Using numerical optimization, quiet zones with other shapes are designed, showing potential for quiet zones with extents that are significantly larger than the well-known limit of a tenth of a wavelength for monopole sources. The paper presents several simulation examples showing quiet zones in various configurations.

  3. Localizing the sources of two independent noises: Role of time varying amplitude differences

    PubMed Central

    Yost, William A.; Brown, Christopher A.

    2013-01-01

    Listeners localized the free-field sources of either one or two simultaneous and independently generated noise bursts. Listeners' localization performance was better when localizing one rather than two sound sources. With two sound sources, localization performance was better when the listener was provided prior information about the location of one of them. Listeners also localized two simultaneous noise bursts that had sinusoidal amplitude modulation (AM) applied, in which the modulation envelope was in-phase across the two source locations or was 180° out-of-phase. The AM was employed to investigate a hypothesis as to what process listeners might use to localize multiple sound sources. The results supported the hypothesis that localization of two sound sources might be based on temporal-spectral regions of the combined waveform in which the sound from one source was more intense than that from the other source. The interaural information extracted from such temporal-spectral regions might provide reliable estimates of the sound source location that produced the more intense sound in that temporal-spectral region. PMID:23556597

  4. Localizing the sources of two independent noises: role of time varying amplitude differences.

    PubMed

    Yost, William A; Brown, Christopher A

    2013-04-01

    Listeners localized the free-field sources of either one or two simultaneous and independently generated noise bursts. Listeners' localization performance was better when localizing one rather than two sound sources. With two sound sources, localization performance was better when the listener was provided prior information about the location of one of them. Listeners also localized two simultaneous noise bursts that had sinusoidal amplitude modulation (AM) applied, in which the modulation envelope was in-phase across the two source locations or was 180° out-of-phase. The AM was employed to investigate a hypothesis as to what process listeners might use to localize multiple sound sources. The results supported the hypothesis that localization of two sound sources might be based on temporal-spectral regions of the combined waveform in which the sound from one source was more intense than that from the other source. The interaural information extracted from such temporal-spectral regions might provide reliable estimates of the sound source location that produced the more intense sound in that temporal-spectral region.

  5. Interactive Sonification of Spontaneous Movement of Children—Cross-Modal Mapping and the Perception of Body Movement Qualities through Sound

    PubMed Central

    Frid, Emma; Bresin, Roberto; Alborno, Paolo; Elblaus, Ludvig

    2016-01-01

    In this paper we present three studies focusing on the effect of different sound models in interactive sonification of bodily movement. We hypothesized that a sound model characterized by continuous smooth sounds would be associated with other movement characteristics than a model characterized by abrupt variation in amplitude and that these associations could be reflected in spontaneous movement characteristics. Three subsequent studies were conducted to investigate the relationship between properties of bodily movement and sound: (1) a motion capture experiment involving interactive sonification of a group of children spontaneously moving in a room, (2) an experiment involving perceptual ratings of sonified movement data and (3) an experiment involving matching between sonified movements and their visualizations in the form of abstract drawings. In (1) we used a system constituting of 17 IR cameras tracking passive reflective markers. The head positions in the horizontal plane of 3–4 children were simultaneously tracked and sonified, producing 3–4 sound sources spatially displayed through an 8-channel loudspeaker system. We analyzed children's spontaneous movement in terms of energy-, smoothness- and directness-index. Despite large inter-participant variability and group-specific effects caused by interaction among children when engaging in the spontaneous movement task, we found a small but significant effect of sound model. Results from (2) indicate that different sound models can be rated differently on a set of motion-related perceptual scales (e.g., expressivity and fluidity). Also, results imply that audio-only stimuli can evoke stronger perceived properties of movement (e.g., energetic, impulsive) than stimuli involving both audio and video representations. Findings in (3) suggest that sounds portraying bodily movement can be represented using abstract drawings in a meaningful way. We argue that the results from these studies support the existence of a cross-modal mapping of body motion qualities from bodily movement to sounds. Sound can be translated and understood from bodily motion, conveyed through sound visualizations in the shape of drawings and translated back from sound visualizations to audio. The work underlines the potential of using interactive sonification to communicate high-level features of human movement data. PMID:27891074

  6. Interactive Sonification of Spontaneous Movement of Children-Cross-Modal Mapping and the Perception of Body Movement Qualities through Sound.

    PubMed

    Frid, Emma; Bresin, Roberto; Alborno, Paolo; Elblaus, Ludvig

    2016-01-01

    In this paper we present three studies focusing on the effect of different sound models in interactive sonification of bodily movement. We hypothesized that a sound model characterized by continuous smooth sounds would be associated with other movement characteristics than a model characterized by abrupt variation in amplitude and that these associations could be reflected in spontaneous movement characteristics. Three subsequent studies were conducted to investigate the relationship between properties of bodily movement and sound: (1) a motion capture experiment involving interactive sonification of a group of children spontaneously moving in a room, (2) an experiment involving perceptual ratings of sonified movement data and (3) an experiment involving matching between sonified movements and their visualizations in the form of abstract drawings. In (1) we used a system constituting of 17 IR cameras tracking passive reflective markers. The head positions in the horizontal plane of 3-4 children were simultaneously tracked and sonified, producing 3-4 sound sources spatially displayed through an 8-channel loudspeaker system. We analyzed children's spontaneous movement in terms of energy-, smoothness- and directness-index. Despite large inter-participant variability and group-specific effects caused by interaction among children when engaging in the spontaneous movement task, we found a small but significant effect of sound model. Results from (2) indicate that different sound models can be rated differently on a set of motion-related perceptual scales (e.g., expressivity and fluidity). Also, results imply that audio-only stimuli can evoke stronger perceived properties of movement (e.g., energetic, impulsive) than stimuli involving both audio and video representations. Findings in (3) suggest that sounds portraying bodily movement can be represented using abstract drawings in a meaningful way. We argue that the results from these studies support the existence of a cross-modal mapping of body motion qualities from bodily movement to sounds. Sound can be translated and understood from bodily motion, conveyed through sound visualizations in the shape of drawings and translated back from sound visualizations to audio. The work underlines the potential of using interactive sonification to communicate high-level features of human movement data.

  7. Sounds Scary? Lack of Habituation following the Presentation of Novel Sounds

    PubMed Central

    Biedenweg, Tine A.; Parsons, Michael H.; Fleming, Patricia A.; Blumstein, Daniel T.

    2011-01-01

    Background Animals typically show less habituation to biologically meaningful sounds than to novel signals. We might therefore expect that acoustic deterrents should be based on natural sounds. Methodology We investigated responses by western grey kangaroos (Macropus fulignosus) towards playback of natural sounds (alarm foot stomps and Australian raven (Corvus coronoides) calls) and artificial sounds (faux snake hiss and bull whip crack). We then increased rate of presentation to examine whether animals would habituate. Finally, we varied frequency of playback to investigate optimal rates of delivery. Principal Findings Nine behaviors clustered into five Principal Components. PC factors 1 and 2 (animals alert or looking, or hopping and moving out of area) accounted for 36% of variance. PC factor 3 (eating cessation, taking flight, movement out of area) accounted for 13% of variance. Factors 4 and 5 (relaxing, grooming and walking; 12 and 11% of variation, respectively) discontinued upon playback. The whip crack was most evocative; eating was reduced from 75% of time spent prior to playback to 6% following playback (post alarm stomp: 32%, raven call: 49%, hiss: 75%). Additionally, 24% of individuals took flight and moved out of area (50 m radius) in response to the whip crack (foot stomp: 0%, raven call: 8% and 4%, hiss: 6%). Increasing rate of presentation (12x/min ×2 min) caused 71% of animals to move out of the area. Conclusions/Significance The bull whip crack, an artificial sound, was as effective as the alarm stomp at eliciting aversive behaviors. Kangaroos did not fully habituate despite hearing the signal up to 20x/min. Highest rates of playback did not elicit the greatest responses, suggesting that ‘more is not always better’. Ultimately, by utilizing both artificial and biological sounds, predictability may be masked or offset, so that habituation is delayed and more effective deterrents may be produced. PMID:21267451

  8. Use of Sound with Digital Text: Moving beyond Sound as an Add-On or Decoration

    ERIC Educational Resources Information Center

    Shanahan, Lynn

    2012-01-01

    The purpose of this interpretive case study was to explore--through a close analysis of one class project--students' use of audio signs and the teacher's scaffolding of the use of audio signs. Two research questions guided this study: (a) In what ways did the fifth-grade students use audio signs, specifically transitions sounds, when constructing…

  9. Illustrations and Supporting Texts for Sound Standing Waves of Air Columns in Pipes in Introductory Physics Textbooks

    ERIC Educational Resources Information Center

    Zeng, Liang; Smith, Chris; Poelzer, G. Herold; Rodriguez, Jennifer; Corpuz, Edgar; Yanev, George

    2014-01-01

    In our pilot studies, we found that many introductory physics textbook illustrations with supporting text for sound standing waves of air columns in open-open, open-closed, and closed-closed pipes inhibit student understanding of sound standing wave phenomena due to student misunderstanding of how air molecules move within these pipes. Based on…

  10. Laboratory Assessment of Commercially Available Ultrasonic Rangefinders

    DTIC Science & Technology

    2015-11-01

    how the room was designed to prevent sound reflections (a combination of the wedges absorbing the waveforms and not having a flat wall ). When testing... sound booth at 0.5 m. ...................................................................................... 5  iv This page is intentionally...environments for sound measurements using a tape measure. This mapping method can be time- consuming and unreliable as objects frequently move around in

  11. Sound-field measurement with moving microphones

    PubMed Central

    Katzberg, Fabrice; Mazur, Radoslaw; Maass, Marco; Koch, Philipp; Mertins, Alfred

    2017-01-01

    Closed-room scenarios are characterized by reverberation, which decreases the performance of applications such as hands-free teleconferencing and multichannel sound reproduction. However, exact knowledge of the sound field inside a volume of interest enables the compensation of room effects and allows for a performance improvement within a wide range of applications. The sampling of sound fields involves the measurement of spatially dependent room impulse responses, where the Nyquist-Shannon sampling theorem applies in the temporal and spatial domains. The spatial measurement often requires a huge number of sampling points and entails other difficulties, such as the need for exact calibration of a large number of microphones. In this paper, a method for measuring sound fields using moving microphones is presented. The number of microphones is customizable, allowing for a tradeoff between hardware effort and measurement time. The goal is to reconstruct room impulse responses on a regular grid from data acquired with microphones between grid positions, in general. For this, the sound field at equidistant positions is related to the measurements taken along the microphone trajectories via spatial interpolation. The benefits of using perfect sequences for excitation, a multigrid recovery, and the prospects for reconstruction by compressed sensing are presented. PMID:28599533

  12. 9- by 15-Foot Low-Speed Wind Tunnel Acoustic Free Field Evaluation; Part 1: Qualification Report Summary and Part 2: Anechoic Environment Characterization Field Test Report

    NASA Technical Reports Server (NTRS)

    Schmitt, Jeff G.; Stahnke, Brian

    2017-01-01

    This report describes test results from an assessment of the acoustically treated 9x15 Foot Low Speed Wind Tunnel at the NASA Glenn Research Center in Cleveland, Ohio in July of 2016. The tests were conducted in accordance with the recently adopted international standard ISO 26101-2012 on Qualification of Free Field Test Environments. This method involves moving a microphone relative to a source and comparing the sound pressure level versus distance measurements with theoretical inverse square law spreading.

  13. Contextual effects on preattentive processing of sound motion as revealed by spatial MMN.

    PubMed

    Shestopalova, L B; Petropavlovskaia, E A; Vaitulevich, S Ph; Nikitin, N I

    2015-04-01

    The magnitude of spatial distance between sound stimuli is critically important for their preattentive discrimination, yet the effect of stimulus context on auditory motion processing is not clear. This study investigated the effects of acoustical change and stimulus context on preattentive spatial change detection. Auditory event-related potentials (ERPs) were recorded for stationary midline noises and two patterns of sound motion produced by linear or abrupt changes of interaural time differences. Each of the three types of stimuli was used as standard or deviant in different blocks. Context effects on mismatch negativity (MMN) elicited by stationary and moving sound stimuli were investigated by reversing the role of standard and deviant stimuli, while the acoustical stimulus parameters were kept the same. That is, MMN amplitudes were calculated by subtracting ERPs to identical stimuli presented as standard in one block and deviant in another block. In contrast, effects of acoustical change on MMN amplitudes were calculated by subtracting ERPs of standards and deviants presented within the same block. Preattentive discrimination of moving and stationary sounds indexed by MMN was strongly dependent on the stimulus context. Higher MMNs were produced in oddball configurations where deviance represented increments of the sound velocity, as compared to configurations with velocity decrements. The effect of standard-deviant reversal was more pronounced with the abrupt sound displacement than with gradual sound motion. Copyright © 2015 Elsevier B.V. All rights reserved.

  14. Automatic analysis of slips of the tongue: Insights into the cognitive architecture of speech production.

    PubMed

    Goldrick, Matthew; Keshet, Joseph; Gustafson, Erin; Heller, Jordana; Needle, Jeremy

    2016-04-01

    Traces of the cognitive mechanisms underlying speaking can be found within subtle variations in how we pronounce sounds. While speech errors have traditionally been seen as categorical substitutions of one sound for another, acoustic/articulatory analyses show they partially reflect the intended sound. When "pig" is mispronounced as "big," the resulting /b/ sound differs from correct productions of "big," moving towards intended "pig"-revealing the role of graded sound representations in speech production. Investigating the origins of such phenomena requires detailed estimation of speech sound distributions; this has been hampered by reliance on subjective, labor-intensive manual annotation. Computational methods can address these issues by providing for objective, automatic measurements. We develop a novel high-precision computational approach, based on a set of machine learning algorithms, for measurement of elicited speech. The algorithms are trained on existing manually labeled data to detect and locate linguistically relevant acoustic properties with high accuracy. Our approach is robust, is designed to handle mis-productions, and overall matches the performance of expert coders. It allows us to analyze a very large dataset of speech errors (containing far more errors than the total in the existing literature), illuminating properties of speech sound distributions previously impossible to reliably observe. We argue that this provides novel evidence that two sources both contribute to deviations in speech errors: planning processes specifying the targets of articulation and articulatory processes specifying the motor movements that execute this plan. These findings illustrate how a much richer picture of speech provides an opportunity to gain novel insights into language processing. Copyright © 2016 Elsevier B.V. All rights reserved.

  15. Sound source localization method in an environment with flow based on Amiet-IMACS

    NASA Astrophysics Data System (ADS)

    Wei, Long; Li, Min; Qin, Sheng; Fu, Qiang; Yang, Debin

    2017-05-01

    A sound source localization method is proposed to localize and analyze the sound source in an environment with airflow. It combines the improved mapping of acoustic correlated sources (IMACS) method and Amiet's method, and is called Amiet-IMACS. It can localize uncorrelated and correlated sound sources with airflow. To implement this approach, Amiet's method is used to correct the sound propagation path in 3D, which improves the accuracy of the array manifold matrix and decreases the position error of the localized source. Then, the mapping of acoustic correlated sources (MACS) method, which is as a high-resolution sound source localization algorithm, is improved by self-adjusting the constraint parameter at each irritation process to increase convergence speed. A sound source localization experiment using a pair of loud speakers in an anechoic wind tunnel under different flow speeds is conducted. The experiment exhibits the advantage of Amiet-IMACS in localizing a more accurate sound source position compared with implementing IMACS alone in an environment with flow. Moreover, the aerodynamic noise produced by a NASA EPPLER 862 STRUT airfoil model in airflow with a velocity of 80 m/s is localized using the proposed method, which further proves its effectiveness in a flow environment. Finally, the relationship between the source position of this airfoil model and its frequency, along with its generation mechanism, is determined and interpreted.

  16. Tracking of "Moving" Fused Auditory Images by Children.

    ERIC Educational Resources Information Center

    Cranford, Jerry L.; And Others

    1993-01-01

    This study evaluated the ability of 30 normally developing children (ages 6-12) to report the perceived location of a stationary fused auditory image (FAI) or track a "moving" FAI. Although subjects performed at normal adult levels with the stationary sound measure, they exhibited a significant age-related trend with the moving sound…

  17. Boundary states at reflective moving boundaries

    NASA Astrophysics Data System (ADS)

    Acosta Minoli, Cesar A.; Kopriva, David A.

    2012-06-01

    We derive and evaluate boundary states for Maxwell's equations, the linear, and the nonlinear Euler gas-dynamics equations to compute wave reflection from moving boundaries. In this study we use a Discontinuous Galerkin Spectral Element method (DGSEM) with Arbitrary Lagrangian-Eulerian (ALE) mapping for the spatial approximation, but the boundary states can be used with other methods, like finite volume schemes. We present four studies using Maxwell's equations, one for the linear Euler equations, and one more for the nonlinear Euler equations. These are: reflection of light from a plane mirror moving at constant velocity, reflection of light from a moving cylinder, reflection of light from a vibrating mirror, reflection of sound from a plane wall and dipole sound generation by an oscillating cylinder in an inviscid flow. The studies show that the boundary states preserve spectral convergence in the solution and in derived quantities like divergence and vorticity.

  18. Modeling and observations of an elevated, moving infrasonic source: Eigenray methods.

    PubMed

    Blom, Philip; Waxler, Roger

    2017-04-01

    The acoustic ray tracing relations are extended by the inclusion of auxiliary parameters describing variations in the spatial ray coordinates and eikonal vector due to changes in the initial conditions. Computation of these parameters allows one to define the geometric spreading factor along individual ray paths and assists in identification of caustic surfaces so that phase shifts can be easily identified. A method is developed leveraging the auxiliary parameters to identify propagation paths connecting specific source-receiver geometries, termed eigenrays. The newly introduced method is found to be highly efficient in cases where propagation is non-planar due to horizontal variations in the propagation medium or the presence of cross winds. The eigenray method is utilized in analysis of infrasonic signals produced by a multi-stage sounding rocket launch with promising results for applications of tracking aeroacoustic sources in the atmosphere and specifically to analysis of motor performance during dynamic tests.

  19. On the horizontal wobbling of an object levitated by near-field acoustic levitation.

    PubMed

    Kim, Cheol-Ho; Ih, Jeong-Guon

    2007-11-01

    A circular planar object can be levitated with several hundreds of microns by ultrasonic near-field acoustic levitation (NFAL). However, when both the sound source and the levitated object are circularly shaped and the center of the levitated object does not coincide with the source center, instability problem often occurs. When this happens, it becomes difficult to pick up or transport the object for the next process. In this study, when the center of the levitated object was offset from the source center, the moving direction of the levitated object was predicted by using the time averaged potential around the levitated object. The wobbling frequency of the levitated object was calculated by analyzing the nonlinear wobbling motion of the object. It was shown that the predicted wobbling frequencies agreed with measured ones well. Finally, a safe zone was suggested to avoid the unstable movement of an object.

  20. Sonorous images through digital holographic images

    NASA Astrophysics Data System (ADS)

    Azevedo, Isabel; Sandford-Richardson, Elizabeth

    2017-03-01

    The art of the last fifty years has significantly surrounded the presence of the body, the relationship between human and interactive technologies. Today in interactive art, there are not only representations that speak of the body but actions and behaviours that involve the body. In holography, the image appears and disappears from the observer's vision field; because the holographic image is light, we can see multidimensional spaces, shapes and colours existing on the same time, presence and absence of the image on the holographic plate. And the image can be flowing in front of the plate that sometimes people try touching it with his hands. That means, to the viewer will be interactive events, with no beginning or end that can be perceived in any direction, forward or backward, depending on the relative position and the time the viewer spends in front of the hologram. To explore that feature we are proposing an installation with four holograms, and several sources of different kind of sounds connected with each hologram. When viewers will move in front of each hologram they will activate different sources of sound. The search is not only about the images in the holograms, but also the looking for different types of sounds that this demand will require. The digital holograms were produced using the HoloCam Portable Light System with the 35 mm camera Canon 700D to capture image information, it was then edited on computer using the Motion 5 and Final Cut Pro X programs.

  1. Fundamentals of diagnostic ultrasonography.

    PubMed

    Noce, J P

    1990-01-01

    Diagnostic ultrasonography uses acoustical waves in the frequency range of 1 to 20 MHz. These waves obey Snell's law of reflection and refraction, which are rules ordinary to wave behavior. In ultrasound, the analogy to momentum is acoustic impedance. The acoustic impedance, Z, is equal to the density, p, times velocity, v. The ultrasound transducer converts electrical energy into ultrasound energy and vice versa. The transducer usually consists of a piezoelectric crystal composed of such ceramic materials as barium titanate, lead titanate, zirconate, or lead metaniobate. Five basic ultrasonic scanning modes play the major roles in clinical applications. A-mode, or amplitude-mode, scanning measures the tissue discontinuity along the scan axis. B-mode scanning produces a two-dimensional image of the tissue under study by combining A-mode signals from various directions through mechanical transducer scanning. M-mode, or time motion scanning, is an extension of the A-mode approach in which a single stationary transducer is used. The depth of the echo is displayed on the vertical axis; the brightness of the oscilloscope display is modulated by the echo amplitude. Real-time scanning, or rapid B-scanning, techniques provide continuous data acquisition at a rate sufficient to give the impression of the instantaneous motion of moving structures. Doppler scanning relies on the presence of motion. The Doppler effect occurs when there is relative motion between the source of sound and the receiver of the sound, causing a change in the detected frequency of the sound source.

  2. Active room compensation for sound reinforcement using sound field separation techniques.

    PubMed

    Heuchel, Franz M; Fernandez-Grande, Efren; Agerkvist, Finn T; Shabalina, Elena

    2018-03-01

    This work investigates how the sound field created by a sound reinforcement system can be controlled at low frequencies. An indoor control method is proposed which actively absorbs the sound incident on a reflecting boundary using an array of secondary sources. The sound field is separated into incident and reflected components by a microphone array close to the secondary sources, enabling the minimization of reflected components by means of optimal signals for the secondary sources. The method is purely feed-forward and assumes constant room conditions. Three different sound field separation techniques for the modeling of the reflections are investigated based on plane wave decomposition, equivalent sources, and the Spatial Fourier transform. Simulations and an experimental validation are presented, showing that the control method performs similarly well at enhancing low frequency responses with the three sound separation techniques. Resonances in the entire room are reduced, although the microphone array and secondary sources are confined to a small region close to the reflecting wall. Unlike previous control methods based on the creation of a plane wave sound field, the investigated method works in arbitrary room geometries and primary source positions.

  3. Analysis of ultrasonically rotating droplet using moving particle semi-implicit and distributed point source methods

    NASA Astrophysics Data System (ADS)

    Wada, Yuji; Yuge, Kohei; Tanaka, Hiroki; Nakamura, Kentaro

    2016-07-01

    Numerical analysis of the rotation of an ultrasonically levitated droplet with a free surface boundary is discussed. The ultrasonically levitated droplet is often reported to rotate owing to the surface tangential component of acoustic radiation force. To observe the torque from an acoustic wave and clarify the mechanism underlying the phenomena, it is effective to take advantage of numerical simulation using the distributed point source method (DPSM) and moving particle semi-implicit (MPS) method, both of which do not require a calculation grid or mesh. In this paper, the numerical treatment of the viscoacoustic torque, which emerges from the viscous boundary layer and governs the acoustical droplet rotation, is discussed. The Reynolds stress traction force is calculated from the DPSM result using the idea of effective normal particle velocity through the boundary layer and input to the MPS surface particles. A droplet levitated in an acoustic chamber is simulated using the proposed calculation method. The droplet is vertically supported by a plane standing wave from an ultrasonic driver and subjected to a rotating sound field excited by two acoustic sources on the side wall with different phases. The rotation of the droplet is successfully reproduced numerically and its acceleration is discussed and compared with those in the literature.

  4. An evaluation of talker localization based on direction of arrival estimation and statistical sound source identification

    NASA Astrophysics Data System (ADS)

    Nishiura, Takanobu; Nakamura, Satoshi

    2002-11-01

    It is very important to capture distant-talking speech for a hands-free speech interface with high quality. A microphone array is an ideal candidate for this purpose. However, this approach requires localizing the target talker. Conventional talker localization algorithms in multiple sound source environments not only have difficulty localizing the multiple sound sources accurately, but also have difficulty localizing the target talker among known multiple sound source positions. To cope with these problems, we propose a new talker localization algorithm consisting of two algorithms. One is DOA (direction of arrival) estimation algorithm for multiple sound source localization based on CSP (cross-power spectrum phase) coefficient addition method. The other is statistical sound source identification algorithm based on GMM (Gaussian mixture model) for localizing the target talker position among localized multiple sound sources. In this paper, we particularly focus on the talker localization performance based on the combination of these two algorithms with a microphone array. We conducted evaluation experiments in real noisy reverberant environments. As a result, we confirmed that multiple sound signals can be identified accurately between ''speech'' or ''non-speech'' by the proposed algorithm. [Work supported by ATR, and MEXT of Japan.

  5. Sound source localization and segregation with internally coupled ears: the treefrog model

    PubMed Central

    Christensen-Dalsgaard, Jakob

    2016-01-01

    Acoustic signaling plays key roles in mediating many of the reproductive and social behaviors of anurans (frogs and toads). Moreover, acoustic signaling often occurs at night, in structurally complex habitats, such as densely vegetated ponds, and in dense breeding choruses characterized by high levels of background noise and acoustic clutter. Fundamental to anuran behavior is the ability of the auditory system to determine accurately the location from where sounds originate in space (sound source localization) and to assign specific sounds in the complex acoustic milieu of a chorus to their correct sources (sound source segregation). Here, we review anatomical, biophysical, neurophysiological, and behavioral studies aimed at identifying how the internally coupled ears of frogs contribute to sound source localization and segregation. Our review focuses on treefrogs in the genus Hyla, as they are the most thoroughly studied frogs in terms of sound source localization and segregation. They also represent promising model systems for future work aimed at understanding better how internally coupled ears contribute to sound source localization and segregation. We conclude our review by enumerating directions for future research on these animals that will require the collaborative efforts of biologists, physicists, and roboticists. PMID:27730384

  6. Application of acoustic radiosity methods to noise propagation within buildings

    NASA Astrophysics Data System (ADS)

    Muehleisen, Ralph T.; Beamer, C. Walter

    2005-09-01

    The prediction of sound pressure levels in rooms from transmitted sound is a difficult problem. The sound energy in the source room incident on the common wall must be accurately predicted. In the receiving room, the propagation of sound from the planar wall source must also be accurately predicted. The radiosity method naturally computes the spatial distribution of sound energy incident on a wall and also naturally predicts the propagation of sound from a planar area source. In this paper, the application of the radiosity method to sound transmission problems is introduced and explained.

  7. Evaluation of moving-coil loudspeaker and passive radiator parameters using normal-incidence sound transmission measurements: theoretical developments.

    PubMed

    Leishman, Timothy W; Anderson, Brian E

    2013-07-01

    The parameters of moving-coil loudspeaker drivers are typically determined using direct electrical excitation and measurement. However, as electro-mechano-acoustical devices, their parameters should also follow from suitable mechanical or acoustical evaluations. This paper presents the theory of an acoustical method of excitation and measurement using normal-incidence sound transmission through a baffled driver as a plane-wave tube partition. Analogous circuits enable key parameters to be extracted from measurement results in terms of open and closed-circuit driver conditions. Associated tools are presented that facilitate adjacent field decompositions and derivations of sound transmission coefficients (in terms of driver parameters) directly from the circuits. The paper also clarifies the impact of nonanechoic receiving tube terminations and the specific benefits of downstream field decompositions.

  8. Ejectable underwater sound source recovery assembly

    NASA Technical Reports Server (NTRS)

    Irick, S. C. (Inventor)

    1974-01-01

    An underwater sound source is described that may be ejectably mounted on any mobile device that travels over water, to facilitate in the location and recovery of the device when submerged. A length of flexible line maintains a connection between the mobile device and the sound source. During recovery, the sound source is located be particularly useful in the recovery of spent rocket motors that bury in the ocean floor upon impact.

  9. Visual Presentation Effects on Identification of Multiple Environmental Sounds

    PubMed Central

    Masakura, Yuko; Ichikawa, Makoto; Shimono, Koichi; Nakatsuka, Reio

    2016-01-01

    This study examined how the contents and timing of a visual stimulus affect the identification of mixed sounds recorded in a daily life environment. For experiments, we presented four environment sounds as auditory stimuli for 5 s along with a picture or a written word as a visual stimulus that might or might not denote the source of one of the four sounds. Three conditions of temporal relations between the visual stimuli and sounds were used. The visual stimulus was presented either: (a) for 5 s simultaneously with the sound; (b) for 5 s, 1 s before the sound (SOA between the audio and visual stimuli was 6 s); or (c) for 33 ms, 1 s before the sound (SOA was 1033 ms). Participants reported all identifiable sounds for those audio–visual stimuli. To characterize the effects of visual stimuli on sound identification, the following were used: the identification rates of sounds for which the visual stimulus denoted its sound source, the rates of other sounds for which the visual stimulus did not denote the sound source, and the frequency of false hearing of a sound that was not presented for each sound set. Results of the four experiments demonstrated that a picture or a written word promoted identification of the sound when it was related to the sound, particularly when the visual stimulus was presented for 5 s simultaneously with the sounds. However, a visual stimulus preceding the sounds had a benefit only for the picture, not for the written word. Furthermore, presentation with a picture denoting a sound simultaneously with the sound reduced the frequency of false hearing. These results suggest three ways that presenting a visual stimulus affects identification of the auditory stimulus. First, activation of the visual representation extracted directly from the picture promotes identification of the denoted sound and suppresses the processing of sounds for which the visual stimulus did not denote the sound source. Second, effects based on processing of the conceptual information promote identification of the denoted sound and suppress the processing of sounds for which the visual stimulus did not denote the sound source. Third, processing of the concurrent visual representation suppresses false hearing. PMID:26973478

  10. The effect of brain lesions on sound localization in complex acoustic environments.

    PubMed

    Zündorf, Ida C; Karnath, Hans-Otto; Lewald, Jörg

    2014-05-01

    Localizing sound sources of interest in cluttered acoustic environments--as in the 'cocktail-party' situation--is one of the most demanding challenges to the human auditory system in everyday life. In this study, stroke patients' ability to localize acoustic targets in a single-source and in a multi-source setup in the free sound field were directly compared. Subsequent voxel-based lesion-behaviour mapping analyses were computed to uncover the brain areas associated with a deficit in localization in the presence of multiple distracter sound sources rather than localization of individually presented sound sources. Analyses revealed a fundamental role of the right planum temporale in this task. The results from the left hemisphere were less straightforward, but suggested an involvement of inferior frontal and pre- and postcentral areas. These areas appear to be particularly involved in the spectrotemporal analyses crucial for effective segregation of multiple sound streams from various locations, beyond the currently known network for localization of isolated sound sources in otherwise silent surroundings.

  11. Source levels of social sounds in migrating humpback whales (Megaptera novaeangliae).

    PubMed

    Dunlop, Rebecca A; Cato, Douglas H; Noad, Michael J; Stokes, Dale M

    2013-07-01

    The source level of an animal sound is important in communication, since it affects the distance over which the sound is audible. Several measurements of source levels of whale sounds have been reported, but the accuracy of many is limited because the distance to the source and the acoustic transmission loss were estimated rather than measured. This paper presents measurements of source levels of social sounds (surface-generated and vocal sounds) of humpback whales from a sample of 998 sounds recorded from 49 migrating humpback whale groups. Sources were localized using a wide baseline five hydrophone array and transmission loss was measured for the site. Social vocalization source levels were found to range from 123 to 183 dB re 1 μPa @ 1 m with a median of 158 dB re 1 μPa @ 1 m. Source levels of surface-generated social sounds ("breaches" and "slaps") were narrower in range (133 to 171 dB re 1 μPa @ 1 m) but slightly higher in level (median of 162 dB re 1 μPa @ 1 m) compared to vocalizations. The data suggest that group composition has an effect on group vocalization source levels in that singletons and mother-calf-singing escort groups tend to vocalize at higher levels compared to other group compositions.

  12. Children's Literature on the Digital Move

    ERIC Educational Resources Information Center

    Wolf, Shelby A.

    2014-01-01

    Children's literature is on the digital move, and many of the new technologies are eye popping in terms of their aesthetic qualities. Books that are already beloved are moving into new spaces well beyond the combination of words and illustration, as movement, sound, and interactivity enter the scene. The purpose of this article is to explore…

  13. Dynamic Spatial Hearing by Human and Robot Listeners

    NASA Astrophysics Data System (ADS)

    Zhong, Xuan

    This study consisted of several related projects on dynamic spatial hearing by both human and robot listeners. The first experiment investigated the maximum number of sound sources that human listeners could localize at the same time. Speech stimuli were presented simultaneously from different loudspeakers at multiple time intervals. The maximum of perceived sound sources was close to four. The second experiment asked whether the amplitude modulation of multiple static sound sources could lead to the perception of auditory motion. On the horizontal and vertical planes, four independent noise sound sources with 60° spacing were amplitude modulated with consecutively larger phase delay. At lower modulation rates, motion could be perceived by human listeners in both cases. The third experiment asked whether several sources at static positions could serve as "acoustic landmarks" to improve the localization of other sources. Four continuous speech sound sources were placed on the horizontal plane with 90° spacing and served as the landmarks. The task was to localize a noise that was played for only three seconds when the listener was passively rotated in a chair in the middle of the loudspeaker array. The human listeners were better able to localize the sound sources with landmarks than without. The other experiments were with the aid of an acoustic manikin in an attempt to fuse binaural recording and motion data to localize sounds sources. A dummy head with recording devices was mounted on top of a rotating chair and motion data was collected. The fourth experiment showed that an Extended Kalman Filter could be used to localize sound sources in a recursive manner. The fifth experiment demonstrated the use of a fitting method for separating multiple sounds sources.

  14. Monitoring CO2 sources and sinks from space : the Orbiting Carbon Observatory (OCO) Mission

    NASA Technical Reports Server (NTRS)

    Crisp, David

    2006-01-01

    NASA's Orbiting Carbon Observatory (OCO) will make the first space-based measurements of atmospheric carbon dioxide (CO2) with the precision, resolution, and coverage needed to characterize the geographic distribution of CO2 sources and sinks and quantify their variability over the seasonal cycle. OCO is currently scheduled for launch in 2008. The observatory will carry a single instrument that incorporates three high-resolution grating spectrometers designed to measure the near-infrared absorption by CO2 and molecular oxygen (O2) in reflected sunlight. OCO will fly 12 minutes ahead of the EOS Aqua platform in the Earth Observing System (EOS) Afternoon Constellation (A-Train). The in-strument will collect 12 to 24 soundings per second as the Observatory moves along its orbit track on the day side of the Earth. A small sampling footprint (<3 km2 at nadir) was adopted to reduce biases in each sounding associated with clouds and aerosols and spatial variations in surface topography. A comprehensive ground-based validation program will be used to assess random errors and biases in the XCO2 product on regional to continental scales. Measurements collected by OCO will be assimilated with other environmental measurements to retrieve surface sources and sinks of CO2. This information could play an important role in monitoring the integrity of large scale CO2 sequestration projects.

  15. Visual motion disambiguation by a subliminal sound.

    PubMed

    Dufour, Andre; Touzalin, Pascale; Moessinger, Michèle; Brochard, Renaud; Després, Olivier

    2008-09-01

    There is growing interest in the effect of sound on visual motion perception. One model involves the illusion created when two identical objects moving towards each other on a two-dimensional visual display can be seen to either bounce off or stream through each other. Previous studies show that the large bias normally seen toward the streaming percept can be modulated by the presentation of an auditory event at the moment of coincidence. However, no reports to date provide sufficient evidence to indicate whether the sound bounce-inducing effect is due to a perceptual binding process or merely to an explicit inference resulting from the transient auditory stimulus resembling a physical collision of two objects. In the present study, we used a novel experimental design in which a subliminal sound was presented either 150 ms before, at, or 150 ms after the moment of coincidence of two disks moving towards each other. The results showed that there was an increased perception of bouncing (rather than streaming) when the subliminal sound was presented at or 150 ms after the moment of coincidence compared to when no sound was presented. These findings provide the first empirical demonstration that activation of the human auditory system without reaching consciousness affects the perception of an ambiguous visual motion display.

  16. Forced sound transmission through a finite-sized single leaf panel subject to a point source excitation.

    PubMed

    Wang, Chong

    2018-03-01

    In the case of a point source in front of a panel, the wavefront of the incident wave is spherical. This paper discusses spherical sound waves transmitting through a finite sized panel. The forced sound transmission performance that predominates in the frequency range below the coincidence frequency is the focus. Given the point source located along the centerline of the panel, forced sound transmission coefficient is derived through introducing the sound radiation impedance for spherical incident waves. It is found that in addition to the panel mass, forced sound transmission loss also depends on the distance from the source to the panel as determined by the radiation impedance. Unlike the case of plane incident waves, sound transmission performance of a finite sized panel does not necessarily converge to that of an infinite panel, especially when the source is away from the panel. For practical applications, the normal incidence sound transmission loss expression of plane incident waves can be used if the distance between the source and panel d and the panel surface area S satisfy d/S>0.5. When d/S ≈0.1, the diffuse field sound transmission loss expression may be a good approximation. An empirical expression for d/S=0  is also given.

  17. 75 FR 45055 - Security Zone; 2010 Seattle Seafair Fleet Week Moving Vessels, Puget Sound, WA

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-08-02

    ... security zones surrounding the HMCS NANAIMO (NCSM 702), HMCS EDMONTON (NCSM 703), and the HMCS BRANDON...), HMCS EDMONTON (NCSM 703), and the HMCS BRANDON (NCSM 710) while underway in the Puget Sound COTP AOR... 703), and the [[Page 45057

  18. Noise Source Identification in a Reverberant Field Using Spherical Beamforming

    NASA Astrophysics Data System (ADS)

    Choi, Young-Chul; Park, Jin-Ho; Yoon, Doo-Byung; Kwon, Hyu-Sang

    Identification of noise sources, their locations and strengths, has been taken great attention. The method that can identify noise sources normally assumes that noise sources are located at a free field. However, the sound in a reverberant field consists of that coming directly from the source plus sound reflected or scattered by the walls or objects in the field. In contrast to the exterior sound field, reflections are added to sound field. Therefore, the source location estimated by the conventional methods may give unacceptable error. In this paper, we explain the effects of reverberant field on interior source identification process and propose the method that can identify noise sources in the reverberant field.

  19. Doppler effect for sound emitted by a moving airborne source and received by acoustic sensors located above and below the sea surface.

    PubMed

    Ferguson, B G

    1993-12-01

    The acoustic emissions from a propeller-driven aircraft are received by a microphone mounted just above ground level and then by a hydrophone located below the sea surface. The dominant feature in the output spectrum of each acoustic sensor is the spectral line corresponding to the propeller blade rate. A frequency estimation technique is applied to the acoustic data from each sensor so that the Doppler shift in the blade rate can be observed at short time intervals during the aircraft's transit overhead. For each acoustic sensor, the observed variation with time of the Doppler-shifted blade rate is compared with the variation predicted by a simple ray-theory model that assumes the atmosphere and the sea are distinct isospeed sound propagation media separated by a plane boundary. The results of the comparison are shown for an aircraft flying with a speed of about 250 kn at altitudes of 500, 700, and 1000 ft.

  20. Coupled auralization and virtual video for immersive multimedia displays

    NASA Astrophysics Data System (ADS)

    Henderson, Paul D.; Torres, Rendell R.; Shimizu, Yasushi; Radke, Richard; Lonsway, Brian

    2003-04-01

    The implementation of maximally-immersive interactive multimedia in exhibit spaces requires not only the presentation of realistic visual imagery but also the creation of a perceptually accurate aural experience. While conventional implementations treat audio and video problems as essentially independent, this research seeks to couple the visual sensory information with dynamic auralization in order to enhance perceptual accuracy. An implemented system has been developed for integrating accurate auralizations with virtual video techniques for both interactive presentation and multi-way communication. The current system utilizes a multi-channel loudspeaker array and real-time signal processing techniques for synthesizing the direct sound, early reflections, and reverberant field excited by a moving sound source whose path may be interactively defined in real-time or derived from coupled video tracking data. In this implementation, any virtual acoustic environment may be synthesized and presented in a perceptually-accurate fashion to many participants over a large listening and viewing area. Subject tests support the hypothesis that the cross-modal coupling of aural and visual displays significantly affects perceptual localization accuracy.

  1. On an integro-differential equation model for the study of the response of an acoustically coupled panel

    NASA Technical Reports Server (NTRS)

    Yen, D. H. Y.; Maestrello, L.; Padula, S.

    1975-01-01

    The response of a clamped panel to supersonically convected turbulence is considered. A theoretical model in the form of an integro-differential equation is employed that takes into account the coupling between the panel motion and the surrounding acoustic medium. The kernels of the integrals, which represent induced pressures due to the panel motion, are Green's functions for sound radiations under various moving and stationary sources. An approximate analysis is made by following a finite-element Ritz-Galerkin procedure. Preliminary numerical results, in agreement with experimental findings, indicate that the acoustic damping is the controlling mechanism of the response.

  2. Bernoulli? Perhaps, but What about Viscosity?

    ERIC Educational Resources Information Center

    Eastwell, Peter

    2007-01-01

    Bernoulli's principle is being misunderstood and consequently misused. This paper clarifies the issues involved, hypothesises as to how this unfortunate situation has arisen, provides sound explanations for many everyday phenomena involving moving air, and makes associated recommendations for teaching the effects of moving fluids.

  3. A Numerical Experiment on the Role of Surface Shear Stress in the Generation of Sound

    NASA Technical Reports Server (NTRS)

    Shariff, Karim; Wang, Meng; Merriam, Marshal (Technical Monitor)

    1996-01-01

    The sound generated due to a localized flow over an infinite flat surface is considered. It is known that the unsteady surface pressure, while appearing in a formal solution to the Lighthill equation, does not constitute a source of sound but rather represents the effect of image quadrupoles. The question of whether a similar surface shear stress term constitutes a true source of dipole sound is less settled. Some have boldly assumed it is a true source while others have argued that, like the surface pressure, it depends on the sound field (via an acoustic boundary layer) and is therefore not a true source. A numerical experiment based on the viscous, compressible Navier-Stokes equations was undertaken to investigate the issue. A small region of a wall was oscillated tangentially. The directly computed sound field was found to to agree with an acoustic analogy based calculation which regards the surface shear as an acoustically compact dipole source of sound.

  4. CLIVAR Mode Water Dynamics Experiment (CLIMODE), Fall 2006 R/V Oceanus Voyage 434, November 16, 2006-December 3, 2006

    DTIC Science & Technology

    2007-12-01

    except for the dive zero time which needed to be programmed during the cruise when the deployment schedule dates were confirmed. _ ACM - Aanderaa ACM...guards bolted on to complete the frame prior to deployment. Sound Source - Sound sources were scheduled to be redeployed. Sound sources were originally...battery voltages and a vacuum. A +27 second time drift was noted and the time was reset. The sound source was scheduled to go to full power on November

  5. Statistics of natural reverberation enable perceptual separation of sound and space

    PubMed Central

    Traer, James; McDermott, Josh H.

    2016-01-01

    In everyday listening, sound reaches our ears directly from a source as well as indirectly via reflections known as reverberation. Reverberation profoundly distorts the sound from a source, yet humans can both identify sound sources and distinguish environments from the resulting sound, via mechanisms that remain unclear. The core computational challenge is that the acoustic signatures of the source and environment are combined in a single signal received by the ear. Here we ask whether our recognition of sound sources and spaces reflects an ability to separate their effects and whether any such separation is enabled by statistical regularities of real-world reverberation. To first determine whether such statistical regularities exist, we measured impulse responses (IRs) of 271 spaces sampled from the distribution encountered by humans during daily life. The sampled spaces were diverse, but their IRs were tightly constrained, exhibiting exponential decay at frequency-dependent rates: Mid frequencies reverberated longest whereas higher and lower frequencies decayed more rapidly, presumably due to absorptive properties of materials and air. To test whether humans leverage these regularities, we manipulated IR decay characteristics in simulated reverberant audio. Listeners could discriminate sound sources and environments from these signals, but their abilities degraded when reverberation characteristics deviated from those of real-world environments. Subjectively, atypical IRs were mistaken for sound sources. The results suggest the brain separates sound into contributions from the source and the environment, constrained by a prior on natural reverberation. This separation process may contribute to robust recognition while providing information about spaces around us. PMID:27834730

  6. Statistics of natural reverberation enable perceptual separation of sound and space.

    PubMed

    Traer, James; McDermott, Josh H

    2016-11-29

    In everyday listening, sound reaches our ears directly from a source as well as indirectly via reflections known as reverberation. Reverberation profoundly distorts the sound from a source, yet humans can both identify sound sources and distinguish environments from the resulting sound, via mechanisms that remain unclear. The core computational challenge is that the acoustic signatures of the source and environment are combined in a single signal received by the ear. Here we ask whether our recognition of sound sources and spaces reflects an ability to separate their effects and whether any such separation is enabled by statistical regularities of real-world reverberation. To first determine whether such statistical regularities exist, we measured impulse responses (IRs) of 271 spaces sampled from the distribution encountered by humans during daily life. The sampled spaces were diverse, but their IRs were tightly constrained, exhibiting exponential decay at frequency-dependent rates: Mid frequencies reverberated longest whereas higher and lower frequencies decayed more rapidly, presumably due to absorptive properties of materials and air. To test whether humans leverage these regularities, we manipulated IR decay characteristics in simulated reverberant audio. Listeners could discriminate sound sources and environments from these signals, but their abilities degraded when reverberation characteristics deviated from those of real-world environments. Subjectively, atypical IRs were mistaken for sound sources. The results suggest the brain separates sound into contributions from the source and the environment, constrained by a prior on natural reverberation. This separation process may contribute to robust recognition while providing information about spaces around us.

  7. The behavior of natural and anthropogenic osmium in Long Island Sound, an urban estuary in the eastern U.S.

    NASA Astrophysics Data System (ADS)

    Williams, Gwyneth; Marcantonio, Franco; Turekian, Karl K.

    1997-04-01

    The Os concentration and 187Os/ 186Os distributions in surface sediments of Long Island Sound (eastern U.S.) provide a way of determining the sources and estuarine transport of Os. The contribution of anthropogenic Os from sewer outfalls from the New York City region supplies a tracer with a characteristic 187Os/ 186Os of about 1. The Os concentration of the bulk surface sediment increases steeply moving toward New York City in the westernmost Sound and generally follows the concentration of organic carbon. The 187Os/ 186Os ratio of bulk surface sediment increases from west to east in the westernmost part of the Sound and is effectively constant in the central Sound. We interpret these results as indicating that the surface bulk sediments of the Sound contain a low 187Os/ 186Os component, perhaps as a reduced coating associated with organic remains from sewer outfalls. The acid hydrogen peroxide leach fraction has an average 187Os/ 186Os of 9.5 in the central Sound, significantly higher than both the bulk sediment value and the probable sea water value of about 8. The leach fraction in the westernmost part of the traverse is less radiogenic than the central Sound and follows the Os wsotope trend of the bulk sediment. Liquid effluent from a New York City sewer outfall contains 30 pg l -1 of dissolved Os with a 187Os/ 186Os of about 2.5, consistent with its being an end-member of the west-east sediment pattern recorded in the leach fractions of the westernmost cores. The leachable Os from the central Sound predominantly reflects Os in ferromanganese oxyhydroxide coatings from continentally derived sediments with 187Os/ 186Os ratios more radiogenic than seawater. The distribution patterns of anthropogenic and natural Os, with their characteristic isotopic signatures in the Sound, and the insights gained from the behavior of other particle-reactive species, indicates that very little Os in solution may pass through the estuarine gauntlet.

  8. Characteristic sounds facilitate visual search.

    PubMed

    Iordanescu, Lucica; Guzman-Martinez, Emmanuel; Grabowecky, Marcia; Suzuki, Satoru

    2008-06-01

    In a natural environment, objects that we look for often make characteristic sounds. A hiding cat may meow, or the keys in the cluttered drawer may jingle when moved. Using a visual search paradigm, we demonstrated that characteristic sounds facilitated visual localization of objects, even when the sounds carried no location information. For example, finding a cat was faster when participants heard a meow sound. In contrast, sounds had no effect when participants searched for names rather than pictures of objects. For example, hearing "meow" did not facilitate localization of the word cat. These results suggest that characteristic sounds cross-modally enhance visual (rather than conceptual) processing of the corresponding objects. Our behavioral demonstration of object-based cross-modal enhancement complements the extensive literature on space-based cross-modal interactions. When looking for your keys next time, you might want to play jingling sounds.

  9. An integrated system for dynamic control of auditory perspective in a multichannel sound field

    NASA Astrophysics Data System (ADS)

    Corey, Jason Andrew

    An integrated system providing dynamic control of sound source azimuth, distance and proximity to a room boundary within a simulated acoustic space is proposed for use in multichannel music and film sound production. The system has been investigated, implemented, and psychoacoustically tested within the ITU-R BS.775 recommended five-channel (3/2) loudspeaker layout. The work brings together physical and perceptual models of room simulation to allow dynamic placement of virtual sound sources at any location of a simulated space within the horizontal plane. The control system incorporates a number of modules including simulated room modes, "fuzzy" sources, and tracking early reflections, whose parameters are dynamically changed according to sound source location within the simulated space. The control functions of the basic elements, derived from theories of perception of a source in a real room, have been carefully tuned to provide efficient, effective, and intuitive control of a sound source's perceived location. Seven formal listening tests were conducted to evaluate the effectiveness of the algorithm design choices. The tests evaluated: (1) loudness calibration of multichannel sound images; (2) the effectiveness of distance control; (3) the resolution of distance control provided by the system; (4) the effectiveness of the proposed system when compared to a commercially available multichannel room simulation system in terms of control of source distance and proximity to a room boundary; (5) the role of tracking early reflection patterns on the perception of sound source distance; (6) the role of tracking early reflection patterns on the perception of lateral phantom images. The listening tests confirm the effectiveness of the system for control of perceived sound source distance, proximity to room boundaries, and azimuth, through fine, dynamic adjustment of parameters according to source location. All of the parameters are grouped and controlled together to create a perceptually strong impression of source location and movement within a simulated space.

  10. Experiments of multichannel least-square methods for sound field reproduction inside aircraft mock-up: Objective evaluations

    NASA Astrophysics Data System (ADS)

    Gauthier, P.-A.; Camier, C.; Lebel, F.-A.; Pasco, Y.; Berry, A.; Langlois, J.; Verron, C.; Guastavino, C.

    2016-08-01

    Sound environment reproduction of various flight conditions in aircraft mock-ups is a valuable tool for the study, prediction, demonstration and jury testing of interior aircraft sound quality and annoyance. To provide a faithful reproduced sound environment, time, frequency and spatial characteristics should be preserved. Physical sound field reproduction methods for spatial sound reproduction are mandatory to immerse the listener's body in the proper sound fields so that localization cues are recreated at the listener's ears. Vehicle mock-ups pose specific problems for sound field reproduction. Confined spaces, needs for invisible sound sources and very specific acoustical environment make the use of open-loop sound field reproduction technologies such as wave field synthesis (based on free-field models of monopole sources) not ideal. In this paper, experiments in an aircraft mock-up with multichannel least-square methods and equalization are reported. The novelty is the actual implementation of sound field reproduction with 3180 transfer paths and trim panel reproduction sources in laboratory conditions with a synthetic target sound field. The paper presents objective evaluations of reproduced sound fields using various metrics as well as sound field extrapolation and sound field characterization.

  11. Modeling of reverberant room responses for two-dimensional spatial sound field analysis and synthesis.

    PubMed

    Bai, Mingsian R; Li, Yi; Chiang, Yi-Hao

    2017-10-01

    A unified framework is proposed for analysis and synthesis of two-dimensional spatial sound field in reverberant environments. In the sound field analysis (SFA) phase, an unbaffled 24-element circular microphone array is utilized to encode the sound field based on the plane-wave decomposition. Depending on the sparsity of the sound sources, the SFA stage can be implemented in two manners. For sparse-source scenarios, a one-stage algorithm based on compressive sensing algorithm is utilized. Alternatively, a two-stage algorithm can be used, where the minimum power distortionless response beamformer is used to localize the sources and Tikhonov regularization algorithm is used to extract the source amplitudes. In the sound field synthesis (SFS), a 32-element rectangular loudspeaker array is employed to decode the target sound field using pressure matching technique. To establish the room response model, as required in the pressure matching step of the SFS phase, an SFA technique for nonsparse-source scenarios is utilized. Choice of regularization parameters is vital to the reproduced sound field. In the SFS phase, three SFS approaches are compared in terms of localization performance and voice reproduction quality. Experimental results obtained in a reverberant room are presented and reveal that an accurate room response model is vital to immersive rendering of the reproduced sound field.

  12. First indications that northern bottlenose whales are sensitive to behavioural disturbance from anthropogenic noise

    PubMed Central

    Miller, P. J. O.; Kvadsheim, P. H.; Lam, F. P. A.; Tyack, P. L.; Curé, C.; DeRuiter, S. L.; Kleivane, L.; Sivle, L. D.; van IJsselmuide, S. P.; Visser, F.; Wensveen, P. J.; von Benda-Beckmann, A. M.; Martín López, L. M.; Narazaki, T.; Hooker, S. K.

    2015-01-01

    Although northern bottlenose whales were the most heavily hunted beaked whale, we have little information about this species in its remote habitat of the North Atlantic Ocean. Underwater anthropogenic noise and disruption of their natural habitat may be major threats, given the sensitivity of other beaked whales to such noise disturbance. We attached dataloggers to 13 northern bottlenose whales and compared their natural sounds and movements to those of one individual exposed to escalating levels of 1–2 kHz upsweep naval sonar signals. At a received sound pressure level (SPL) of 98 dB re 1 μPa, the whale turned to approach the sound source, but at a received SPL of 107 dB re 1 μPa, the whale began moving in an unusually straight course and then made a near 180° turn away from the source, and performed the longest and deepest dive (94 min, 2339 m) recorded for this species. Animal movement parameters differed significantly from baseline for more than 7 h until the tag fell off 33–36 km away. No clicks were emitted during the response period, indicating cessation of normal echolocation-based foraging. A sharp decline in both acoustic and visual detections of conspecifics after exposure suggests other whales in the area responded similarly. Though more data are needed, our results indicate high sensitivity of this species to acoustic disturbance, with consequent risk from marine industrialization and naval activity. PMID:26543576

  13. Sound Source Localization and Speech Understanding in Complex Listening Environments by Single-sided Deaf Listeners After Cochlear Implantation.

    PubMed

    Zeitler, Daniel M; Dorman, Michael F; Natale, Sarah J; Loiselle, Louise; Yost, William A; Gifford, Rene H

    2015-09-01

    To assess improvements in sound source localization and speech understanding in complex listening environments after unilateral cochlear implantation for single-sided deafness (SSD). Nonrandomized, open, prospective case series. Tertiary referral center. Nine subjects with a unilateral cochlear implant (CI) for SSD (SSD-CI) were tested. Reference groups for the task of sound source localization included young (n = 45) and older (n = 12) normal-hearing (NH) subjects and 27 bilateral CI (BCI) subjects. Unilateral cochlear implantation. Sound source localization was tested with 13 loudspeakers in a 180 arc in front of the subject. Speech understanding was tested with the subject seated in an 8-loudspeaker sound system arrayed in a 360-degree pattern. Directionally appropriate noise, originally recorded in a restaurant, was played from each loudspeaker. Speech understanding in noise was tested using the Azbio sentence test and sound source localization quantified using root mean square error. All CI subjects showed poorer-than-normal sound source localization. SSD-CI subjects showed a bimodal distribution of scores: six subjects had scores near the mean of those obtained by BCI subjects, whereas three had scores just outside the 95th percentile of NH listeners. Speech understanding improved significantly in the restaurant environment when the signal was presented to the side of the CI. Cochlear implantation for SSD can offer improved speech understanding in complex listening environments and improved sound source localization in both children and adults. On tasks of sound source localization, SSD-CI patients typically perform as well as BCI patients and, in some cases, achieve scores at the upper boundary of normal performance.

  14. Position-dependent hearing in three species of bushcrickets (Tettigoniidae, Orthoptera)

    PubMed Central

    Lakes-Harlan, Reinhard; Scherberich, Jan

    2015-01-01

    A primary task of auditory systems is the localization of sound sources in space. Sound source localization in azimuth is usually based on temporal or intensity differences of sounds between the bilaterally arranged ears. In mammals, localization in elevation is possible by transfer functions at the ear, especially the pinnae. Although insects are able to locate sound sources, little attention is given to the mechanisms of acoustic orientation to elevated positions. Here we comparatively analyse the peripheral hearing thresholds of three species of bushcrickets in respect to sound source positions in space. The hearing thresholds across frequencies depend on the location of a sound source in the three-dimensional hearing space in front of the animal. Thresholds differ for different azimuthal positions and for different positions in elevation. This position-dependent frequency tuning is species specific. Largest differences in thresholds between positions are found in Ancylecha fenestrata. Correspondingly, A. fenestrata has a rather complex ear morphology including cuticular folds covering the anterior tympanal membrane. The position-dependent tuning might contribute to sound source localization in the habitats. Acoustic orientation might be a selective factor for the evolution of morphological structures at the bushcricket ear and, speculatively, even for frequency fractioning in the ear. PMID:26543574

  15. Position-dependent hearing in three species of bushcrickets (Tettigoniidae, Orthoptera).

    PubMed

    Lakes-Harlan, Reinhard; Scherberich, Jan

    2015-06-01

    A primary task of auditory systems is the localization of sound sources in space. Sound source localization in azimuth is usually based on temporal or intensity differences of sounds between the bilaterally arranged ears. In mammals, localization in elevation is possible by transfer functions at the ear, especially the pinnae. Although insects are able to locate sound sources, little attention is given to the mechanisms of acoustic orientation to elevated positions. Here we comparatively analyse the peripheral hearing thresholds of three species of bushcrickets in respect to sound source positions in space. The hearing thresholds across frequencies depend on the location of a sound source in the three-dimensional hearing space in front of the animal. Thresholds differ for different azimuthal positions and for different positions in elevation. This position-dependent frequency tuning is species specific. Largest differences in thresholds between positions are found in Ancylecha fenestrata. Correspondingly, A. fenestrata has a rather complex ear morphology including cuticular folds covering the anterior tympanal membrane. The position-dependent tuning might contribute to sound source localization in the habitats. Acoustic orientation might be a selective factor for the evolution of morphological structures at the bushcricket ear and, speculatively, even for frequency fractioning in the ear.

  16. Computational study of the interaction between a shock and a near-wall vortex using a weighted compact nonlinear scheme

    NASA Astrophysics Data System (ADS)

    Zuo, Zhifeng; Maekawa, Hiroshi

    2014-02-01

    The interaction between a moderate-strength shock wave and a near-wall vortex is studied numerically by solving the two-dimensional, unsteady compressible Navier-Stokes equations using a weighted compact nonlinear scheme with a simple low-dissipation advection upstream splitting method for flux splitting. Our main purpose is to clarify the development of the flow field and the generation of sound waves resulting from the interaction. The effects of the vortex-wall distance on the sound generation associated with variations in the flow structures are also examined. The computational results show that three sound sources are involved in this problem: (i) a quadrupolar sound source due to the shock-vortex interaction; (ii) a dipolar sound source due to the vortex-wall interaction; and (iii) a dipolar sound source due to unsteady wall shear stress. The sound field is the combination of the sound waves produced by all three sound sources. In addition to the interaction of the incident shock with the vortex, a secondary shock-vortex interaction is caused by the reflection of the reflected shock (MR2) from the wall. The flow field is dominated by the primary and secondary shock-vortex interactions. The generation mechanism of the third sound, which is newly discovered, due to the MR2-vortex interaction is presented. The pressure variations generated by (ii) become significant with decreasing vortex-wall distance. The sound waves caused by (iii) are extremely weak compared with those caused by (i) and (ii) and are negligible in the computed sound field.

  17. Ultrasound sounding in air by fast-moving receiver

    NASA Astrophysics Data System (ADS)

    Sukhanov, D.; Erzakova, N.

    2018-05-01

    A method of ultrasound imaging in the air for a fast receiver. The case, when the speed of movement of the receiver can not be neglected with respect to the speed of sound. In this case, the Doppler effect is significant, making it difficult for matched filtering of the backscattered signal. The proposed method does not use a continuous repetitive noise-sounding signal. generalized approach applies spatial matched filtering in the time domain to recover the ultrasonic tomographic images.

  18. Detection and modeling of the acoustic perturbation produced by the launch of the Space Shuttle using the Global Positioning System

    NASA Astrophysics Data System (ADS)

    Bowling, T. J.; Calais, E.; Dautermann, T.

    2010-12-01

    Rocket launches are known to produce infrasonic pressure waves that propagate into the ionosphere where coupling between electrons and neutral particles induces fluctuations in ionospheric electron density observable in GPS measurements. We have detected ionospheric perturbations following the launch of space shuttle Atlantis on 11 May 2009 using an array of continually operating GPS stations across the Southeastern coast of the United States and in the Caribbean. Detections are prominent to the south of the westward shuttle trajectory in the area of maximum coupling between the acoustic wave and Earth’s magnetic field, move at speeds consistent with the speed of sound, and show coherency between stations covering a large geographic range. We model the perturbation as an explosive source located at the point of closest approach between the shuttle path and each sub-ionospheric point. The neutral pressure wave is propagated using ray tracing, resultant changes in electron density are calculated at points of intersection between rays and satellite-to-reciever line-of-sight, and synthetic integrated electron content values are derived. Arrival times of the observed and synthesized waveforms match closely, with discrepancies related to errors in the apriori sound speed model used for ray tracing. Current work includes the estimation of source location and energy.

  19. Characterization of the Reverberation Chamber at the NASA Langley Structural Acoustics Loads and Transmission (SALT) Facility

    NASA Technical Reports Server (NTRS)

    Grosveld, Ferdinand W.

    2013-01-01

    In 2011 the noise generating capabilities in the reverberation chamber of the Structural Acoustic Loads and Transmission (SALT) facility at NASA Langley Research Center were enhanced with two fiberglass reinforced polyester resin exponential horns, each coupled to Wyle Acoustic Source WAS-3000 airstream modulators. This report describes the characterization of the reverberation chamber in terms of the background noise, diffusivity, sound pressure levels, the reverberation times and the related overall acoustic absorption in the empty chamber and with the acoustic horn(s) installed. The frequency range of interest includes the 80 Hz to 8000 Hz one-third octave bands. Reverberation time and sound pressure level measurements were conducted and standard deviations from the mean were computed. It was concluded that a diffuse field could be produced above the Schroeder frequency in the 400 Hz one-third octave band and higher for all applications. This frequency could be lowered by installing panel diffusers or moving vanes to improve the acoustic modal overlap in the chamber. In the 80 Hz to 400 Hz one-third octave bands a successful measurement will be dependent on the type of measurement, the test configuration, the source and microphone locations and the desired accuracy. It is recommended that qualification measurements endorsed in the International Standards be conducted for each particular application.

  20. Evaluation of small arms noise in a natural soundscape-Bear Butte, SD

    NASA Astrophysics Data System (ADS)

    Braslau, David

    2005-09-01

    Most studies on soundscape intrusion have been limited to moving sources. Of less concern is noise from small arms. Potential impact was predicted from a proposed large small arms facility with 10000 rounds or more per day on the natural soundscape at Bear Butte, one of the most sacred sites of the Northern Cheyenne and other tribes. The primary impacted activity is meditation and oneness with the natural environment that can continue for several days through day and night. Non-natural sources included limited vehicles on a nearby highway and farm equipment, but few aircraft. Second-by-second ambient octave band readings were taken at 20 sites starting before sunrise. The minimum ambient level observed was 19.6 dBA but limits were encountered with a 1/2 in. microphone. Sound level data on small arms were projected from the proposed range four miles north of Bear Butte to elevated points on the Butte assuming a zero wind environment. Impact was evaluated using audibility, intrusiveness and impulse-weighted DNL. Projected levels were well above ambient. While the DNL was projected to increase by 15 dBA, this metric has little meaning for this type of activity. Assumptions related to outdoor sound propagation, audibility and impulsive noise perception are discussed.

  1. J-85 jet engine noise measured in the ONERA S1 wind tunnel and extrapolated to far field

    NASA Technical Reports Server (NTRS)

    Soderman, Paul T.; Julienne, Alain; Atencio, Adolph, Jr.

    1991-01-01

    Noise from a J-85 turbojet with a conical, convergent nozzle was measured in simulated flight in the ONERA S1 Wind Tunnel. Data are presented for several flight speeds up to 130 m/sec and for radiation angles of 40 to 160 degrees relative to the upstream direction. The jet was operated with subsonic and sonic exhaust speeds. A moving microphone on a 2 m sideline was used to survey the radiated sound field in the acoustically treated, closed test section. The data were extrapolated to a 122 m sideline by means of a multiple-sideline source-location method, which was used to identify the acoustic source regions, directivity patterns, and near field effects. The source-location method is described along with its advantages and disadvantages. Results indicate that the effects of simulated flight on J-85 noise are significant. At the maximum forward speed of 130 m/sec, the peak overall sound levels in the aft quadrant were attentuated approximately 10 dB relative to sound levels of the engine operated statically. As expected, the simulated flight and static data tended to merge in the forward quadrant as the radiation angle approached 40 degrees. There is evidence that internal engine or shock noise was important in the forward quadrant. The data are compared with published predictions for flight effects on pure jet noise and internal engine noise. A new empirical prediction is presented that relates the variation of internally generated engine noise or broadband shock noise to forward speed. Measured near field noise extrapolated to far field agrees reasonably well with data from similar engines tested statically outdoors, in flyover, in a wind tunnel, and on the Bertin Aerotrain. Anomalies in the results for the forward quadrant and for angles above 140 degrees are discussed. The multiple-sideline method proved to be cumbersome in this application, and it did not resolve all of the uncertainties associated with measurements of jet noise close to the jet. The simulation was complicated by wind-tunnel background noise and the propagation of low frequency sound around the circuit.

  2. Articulatory speech synthesis and speech production modelling

    NASA Astrophysics Data System (ADS)

    Huang, Jun

    This dissertation addresses the problem of speech synthesis and speech production modelling based on the fundamental principles of human speech production. Unlike the conventional source-filter model, which assumes the independence of the excitation and the acoustic filter, we treat the entire vocal apparatus as one system consisting of a fluid dynamic aspect and a mechanical part. We model the vocal tract by a three-dimensional moving geometry. We also model the sound propagation inside the vocal apparatus as a three-dimensional nonplane-wave propagation inside a viscous fluid described by Navier-Stokes equations. In our work, we first propose a combined minimum energy and minimum jerk criterion to estimate the dynamic vocal tract movements during speech production. Both theoretical error bound analysis and experimental results show that this method can achieve very close match at the target points and avoid the abrupt change in articulatory trajectory at the same time. Second, a mechanical vocal fold model is used to compute the excitation signal of the vocal tract. The advantage of this model is that it is closely coupled with the vocal tract system based on fundamental aerodynamics. As a result, we can obtain an excitation signal with much more detail than the conventional parametric vocal fold excitation model. Furthermore, strong evidence of source-tract interaction is observed. Finally, we propose a computational model of the fricative and stop types of sounds based on the physical principles of speech production. The advantage of this model is that it uses an exogenous process to model the additional nonsteady and nonlinear effects due to the flow mode, which are ignored by the conventional source- filter speech production model. A recursive algorithm is used to estimate the model parameters. Experimental results show that this model is able to synthesize good quality fricative and stop types of sounds. Based on our dissertation work, we carefully argue that the articulatory speech production model has the potential to flexibly synthesize natural-quality speech sounds and to provide a compact computational model for speech production that can be beneficial to a wide range of areas in speech signal processing.

  3. Temporal-Spectral Characterization and Classification of Marine Mammal Vocalizations and Diesel-Electric Ships Radiated Sound over Continental Shelf Scale Regions with Coherent Hydrophone Array Measurements

    NASA Astrophysics Data System (ADS)

    Huang, Wei

    The passive ocean acoustic waveguide remote sensing (POAWRS) technology is capable of monitoring a large variety of underwater sound sources over instantaneous wide areas spanning continental-shelf scale regions. POAWRS uses a large-aperture densely-sampled coherent hydrophone array to significantly enhance the signal-to-noise ratio via beamforming, enabling detection of sound sources roughly two-orders of magnitude more distant in range than that possible with a single hydrophone. The sound sources detected by POAWRS include ocean biology, geophysical processes, and man-made activities. POAWRS provides detection, bearing-time estimation, localization, and classification of underwater sound sources. The volume of underwater sounds detected by POAWRS is immense, typically exceeding a million unique signal detections per day, in the 10-4000 Hz frequency range, making it a tremendously challenging task to distinguish and categorize the various sound sources present in a given region. Here we develop various approaches for characterizing and clustering the signal detections for various subsets of data acquired using the POAWRS technology. The approaches include pitch tracking of the dominant signal detections, time-frequency feature extraction, clustering and categorization methods. These approaches are essential for automatic processing and enhancing the efficiency and accuracy of POAWRS data analysis. The results of the signal detection, clustering and classification analysis are required for further POAWRS processing, including localization and tracking of a large number of oceanic sound sources. Here the POAWRS detection, localization and clustering approaches are applied to analyze and elucidate the vocalization behavior of humpback, sperm and fin whales in the New England continental shelf and slope, including the Gulf of Maine from data acquired using coherent hydrophone arrays. The POAWRS technology can also be applied for monitoring ocean vehicles. Here the approach is calibrated by application to known ships present in the Gulf of Maine and in the Norwegian Sea from their underwater sounds received using a coherent hydrophone array. The vocalization behavior of humpback whales was monitored over vast areas of the Gulf of Maine using the POAWRS technique over multiple diel cycles in Fall 2006. The humpback vocalizations, received at a rate of roughly 1800+/-1100 calls per day, comprised of both song and non-song. The song vocalizations, composed of highly structured and repeatable set of phrases, are characterized by inter-pulse intervals of 3.5 +/- 1.8 s. Songs were detected throughout the diel cycle, occuring roughly 40% during the day and 60% during the night. The humpback non-song vocalizations, dominated by shorter duration (≤3 s) downsweep and bow-shaped moans, as well as a small fraction of longer duration (˜5 s) cries, have significantly larger mean and more variable inter-pulse intervals of 14.2 +/- 11 s. The non-song vocalizations were detected at night with negligible detections during the day, implying they probably function as nighttime communication signals. The humpback song and non-song vocalizations are separately localized using the moving array triangulation and array invariant techniques. The humpback song and non-song moan calls are both consistently localized to a dense area on northeastern Georges Bank and a less dense region extended from Franklin Basin to the Great South Channel. Humpback cries occur exclusively on northeastern Georges Bank and during nights with coincident dense Atlantic herring shoaling populations, implying the cries are feeding-related. Sperm whales in the New England continental shelf and slope were passively localized and classified from their vocalizations received using a single low-frequency (<2500 Hz) densely-sampled horizontal coherent hydrophone array deployed in Spring 2013 in Gulf of Maine. Whale bearings were estimated using time-domain beamforming that provided high coherent array gain in sperm whale click signal-to-noise ratio. Whale ranges from the receiver array center were estimated using the moving array triangulation technique from a sequence of whale bearing measurements. Multiple concurrently vocalizing sperm whales, in the far-field of the horizontal receiver array, were distinguished and classified based on their horizontal spatial locations and the inter-pulse intervals of their vocalized click signals. We provide detailed analysis of over 15,000 fin whale 20 Hz vocalizations received on Oct 1-3, 2006 in the Gulf of Maine. These vocalizations are separated into 16 clusters following the clustering approaches. Seven of these types are prominent, each acounting for between 8% to 16% and together comprise roughly 85% of all the analyzed vocalizations. The 7 prominent clusters are each more abundant during nighttime hours by a factor of roughly 2.5 times than that of the daytime. The diel-spatial correlation of the 7 prominent clusters to the simultaneously observed densities of their fish prey, the Atlantic herring in the Gulf of Maine, is provided which implies that the factor of roughly 2.5 increase in call rate during night-time hours can be attributed to increased fish-feeding activities. (Abstract shortened by ProQuest.).

  4. Reconstruction of sound source signal by analytical passive TR in the environment with airflow

    NASA Astrophysics Data System (ADS)

    Wei, Long; Li, Min; Yang, Debin; Niu, Feng; Zeng, Wu

    2017-03-01

    In the acoustic design of air vehicles, the time-domain signals of noise sources on the surface of air vehicles can serve as data support to reveal the noise source generation mechanism, analyze acoustic fatigue, and take measures for noise insulation and reduction. To rapidly reconstruct the time-domain sound source signals in an environment with flow, a method combining the analytical passive time reversal mirror (AP-TR) with a shear flow correction is proposed. In this method, the negative influence of flow on sound wave propagation is suppressed by the shear flow correction, obtaining the corrected acoustic propagation time delay and path. Those corrected time delay and path together with the microphone array signals are then submitted to the AP-TR, reconstructing more accurate sound source signals in the environment with airflow. As an analytical method, AP-TR offers a supplementary way in 3D space to reconstruct the signal of sound source in the environment with airflow instead of the numerical TR. Experiments on the reconstruction of the sound source signals of a pair of loud speakers are conducted in an anechoic wind tunnel with subsonic airflow to validate the effectiveness and priorities of the proposed method. Moreover the comparison by theorem and experiment result between the AP-TR and the time-domain beamforming in reconstructing the sound source signal is also discussed.

  5. Localization of sound sources in a room with one microphone

    NASA Astrophysics Data System (ADS)

    Peić Tukuljac, Helena; Lissek, Hervé; Vandergheynst, Pierre

    2017-08-01

    Estimation of the location of sound sources is usually done using microphone arrays. Such settings provide an environment where we know the difference between the received signals among different microphones in the terms of phase or attenuation, which enables localization of the sound sources. In our solution we exploit the properties of the room transfer function in order to localize a sound source inside a room with only one microphone. The shape of the room and the position of the microphone are assumed to be known. The design guidelines and limitations of the sensing matrix are given. Implementation is based on the sparsity in the terms of voxels in a room that are occupied by a source. What is especially interesting about our solution is that we provide localization of the sound sources not only in the horizontal plane, but in the terms of the 3D coordinates inside the room.

  6. Hearing Scenes: A Neuromagnetic Signature of Auditory Source and Reverberant Space Separation

    PubMed Central

    Oliva, Aude

    2017-01-01

    Abstract Perceiving the geometry of surrounding space is a multisensory process, crucial to contextualizing object perception and guiding navigation behavior. Humans can make judgments about surrounding spaces from reverberation cues, caused by sounds reflecting off multiple interior surfaces. However, it remains unclear how the brain represents reverberant spaces separately from sound sources. Here, we report separable neural signatures of auditory space and source perception during magnetoencephalography (MEG) recording as subjects listened to brief sounds convolved with monaural room impulse responses (RIRs). The decoding signature of sound sources began at 57 ms after stimulus onset and peaked at 130 ms, while space decoding started at 138 ms and peaked at 386 ms. Importantly, these neuromagnetic responses were readily dissociable in form and time: while sound source decoding exhibited an early and transient response, the neural signature of space was sustained and independent of the original source that produced it. The reverberant space response was robust to variations in sound source, and vice versa, indicating a generalized response not tied to specific source-space combinations. These results provide the first neuromagnetic evidence for robust, dissociable auditory source and reverberant space representations in the human brain and reveal the temporal dynamics of how auditory scene analysis extracts percepts from complex naturalistic auditory signals. PMID:28451630

  7. Converting a Monopole Emission into a Dipole Using a Subwavelength Structure

    NASA Astrophysics Data System (ADS)

    Fan, Xu-Dong; Zhu, Yi-Fan; Liang, Bin; Cheng, Jian-chun; Zhang, Likun

    2018-03-01

    High-efficiency emission of multipoles is unachievable by a source much smaller than the wavelength, preventing compact acoustic devices for generating directional sound beams. Here, we present a primary scheme towards solving this problem by numerically and experimentally enclosing a monopole sound source in a structure with a dimension of around 1 /10 sound wavelength to emit a dipolar field. The radiated sound power is found to be more than twice that of a bare dipole. Our study of efficient emission of directional low-frequency sound from a monopole source in a subwavelength space may have applications such as focused ultrasound for imaging, directional underwater sound beams, miniaturized sonar, etc.

  8. Characteristic sounds facilitate visual search

    PubMed Central

    Iordanescu, Lucica; Guzman-Martinez, Emmanuel; Grabowecky, Marcia; Suzuki, Satoru

    2009-01-01

    In a natural environment, objects that we look for often make characteristic sounds. A hiding cat may meow, or the keys in the cluttered drawer may jingle when moved. Using a visual search paradigm, we demonstrated that characteristic sounds facilitated visual localization of objects, even when the sounds carried no location information. For example, finding a cat was faster when participants heard a meow sound. In contrast, sounds had no effect when participants searched for names rather than pictures of objects. For example, hearing “meow” did not facilitate localization of the word cat. These results suggest that characteristic sounds cross-modally enhance visual (rather than conceptual) processing of the corresponding objects. Our behavioral demonstration of object-based cross-modal enhancement complements the extensive literature on space-based cross-modal interactions. When looking for your keys next time, you might want to play jingling sounds. PMID:18567253

  9. Effect of the spectrum of a high-intensity sound source on the sound-absorbing properties of a resonance-type acoustic lining

    NASA Astrophysics Data System (ADS)

    Ipatov, M. S.; Ostroumov, M. N.; Sobolev, A. F.

    2012-07-01

    Experimental results are presented on the effect of both the sound pressure level and the type of spectrum of a sound source on the impedance of an acoustic lining. The spectra under study include those of white noise, a narrow-band signal, and a signal with a preset waveform. It is found that, to obtain reliable data on the impedance of an acoustic lining from the results of interferometric measurements, the total sound pressure level of white noise or the maximal sound pressure level of a pure tone (at every oscillation frequency) needs to be identical to the total sound pressure level of the actual source at the site of acoustic lining on the channel wall.

  10. 3D Sound Techniques for Sound Source Elevation in a Loudspeaker Listening Environment

    NASA Astrophysics Data System (ADS)

    Kim, Yong Guk; Jo, Sungdong; Kim, Hong Kook; Jang, Sei-Jin; Lee, Seok-Pil

    In this paper, we propose several 3D sound techniques for sound source elevation in stereo loudspeaker listening environments. The proposed method integrates a head-related transfer function (HRTF) for sound positioning and early reflection for adding reverberant circumstance. In addition, spectral notch filtering and directional band boosting techniques are also included for increasing elevation perception capability. In order to evaluate the elevation performance of the proposed method, subjective listening tests are conducted using several kinds of sound sources such as white noise, sound effects, speech, and music samples. It is shown from the tests that the degrees of perceived elevation by the proposed method are around the 17º to 21º when the stereo loudspeakers are located on the horizontal plane.

  11. An experimental investigation of thermoacoustic lasers operating in audible frequency range

    NASA Astrophysics Data System (ADS)

    Kolhe, Sanket Anil

    Thermoacoustic lasers convert heat from a high-temperature heat source into acoustic power while rejecting waste heat to a low temperature sink. The working fluids involved can be air or noble gases which are nontoxic and environmentally benign. Simple in construction due to absence of moving parts, thermoacoustic lasers can be employed to achieve generation of electricity at individual homes, water-heating for domestic purposes, and to facilitate space heating and cooling. The possibility of utilizing waste heat or solar energy to run thermoacoustic devices makes them technically promising and economically viable to generate large quantities of acoustic energy. The research presented in this thesis deals with the effects of geometric parameters (stack position, stack length, tube length) associated with a thermoacoustic laser on the output sound wave. The effects of varying input power on acoustic output were also studied. Based on the experiments, optimum operating conditions were identified and qualitative and/or quantitative explanations were provided to justify our observations. It was observed that the maximum sound pressure level was generated for the laser with the stack positioned at a distance of quarter lengths of a resonator from the closed end. Higher sound pressure levels were recorded for the laser with longer stack lengths and longer resonator lengths. Efforts were also made to develop high-frequency thermoacoustic lasers.

  12. A multi-band spectral subtraction-based algorithm for real-time noise cancellation applied to gunshot acoustics

    NASA Astrophysics Data System (ADS)

    Ramos, António L. L.; Holm, Sverre; Gudvangen, Sigmund; Otterlei, Ragnvald

    2013-06-01

    Acoustical sniper positioning is based on the detection and direction-of-arrival estimation of the shockwave and the muzzle blast acoustical signals. In real-life situations, the detection and direction-of-arrival estimation processes is usually performed under the influence of background noise sources, e.g., vehicles noise, and might result in non-negligible inaccuracies than can affect the system performance and reliability negatively, specially when detecting the muzzle sound under long range distance and absorbing terrains. This paper introduces a multi-band spectral subtraction based algorithm for real-time noise reduction, applied to gunshot acoustical signals. The ballistic shockwave and the muzzle blast signals exhibit distinct frequency contents that are affected differently by additive noise. In most real situations, the noise component is colored and a multi-band spectral subtraction approach for noise reduction contributes to reducing the presence of artifacts in denoised signals. The proposed algorithm is tested using a dataset generated by combining signals from real gunshots and real vehicle noise. The noise component was generated using a steel tracked military tank running on asphalt and includes, therefore, the sound from the vehicle engine, which varies slightly in frequency over time according to the engine's rpm, and the sound from the steel tracks as the vehicle moves.

  13. 30 CFR 57.14200 - Warnings prior to starting or moving equipment.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 30 Mineral Resources 1 2011-07-01 2011-07-01 false Warnings prior to starting or moving equipment... NONMETAL MINES Machinery and Equipment Safety Practices and Operational Procedures § 57.14200 Warnings... equipment, equipment operators shall sound a warning that is audible above the surrounding noise level or...

  14. 30 CFR 57.14200 - Warnings prior to starting or moving equipment.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 30 Mineral Resources 1 2013-07-01 2013-07-01 false Warnings prior to starting or moving equipment... NONMETAL MINES Machinery and Equipment Safety Practices and Operational Procedures § 57.14200 Warnings... equipment, equipment operators shall sound a warning that is audible above the surrounding noise level or...

  15. Recent paleoseismicity record in Prince William Sound, Alaska, USA

    NASA Astrophysics Data System (ADS)

    Kuehl, Steven A.; Miller, Eric J.; Marshall, Nicole R.; Dellapenna, Timothy M.

    2017-12-01

    Sedimentological and geochemical investigation of sediment cores collected in the deep (>400 m) central basin of Prince William Sound, along with geochemical fingerprinting of sediment source areas, are used to identify earthquake-generated sediment gravity flows. Prince William Sound receives sediment from two distinct sources: from offshore (primarily Copper River) through Hinchinbrook Inlet, and from sources within the Sound (primarily Columbia Glacier). These sources are found to have diagnostic elemental ratios indicative of provenance; Copper River Basin sediments were significantly higher in Sr/Pb and Cu/Pb, whereas Prince William Sound sediments were significantly higher in K/Ca and Rb/Sr. Within the past century, sediment gravity flows deposited within the deep central channel of Prince William Sound have robust geochemical (provenance) signatures that can be correlated with known moderate to large earthquakes in the region. Given the thick Holocene sequence in the Sound ( 200 m) and correspondingly high sedimentation rates (>1 cm year-1), this relationship suggests that sediments within the central basin of Prince William Sound may contain an extraordinary high-resolution record of paleoseismicity in the region.

  16. The assessment and evaluation of low-frequency noise near the region of infrasound.

    PubMed

    Ziaran, Stanislav

    2014-01-01

    The main aim of this paper is to present recent knowledge about the assessment and evaluation of low-frequency sounds (noise) and infrasound, close to the threshold of hearing, and identify their potential effect on human health and annoyance. Low-frequency noise generated by air flowing over a moving car with an open window was chosen as a typical scenario which can be subjectively assessed by people traveling by automobile. The principle of noise generated within the interior of the car and its effects on the comfort of the driver and passengers are analyzed at different velocities. An open window of a car at high velocity behaves as a source of specifically strong tonal low-frequency noise which is generally perceived as annoying. The interior noise generated by an open window of a passenger car was measured under different conditions: Driving on a highway and driving on a typical roadway. First, an octave-band analysis was used to assess the noise level and its impact on the driver's comfort. Second, a fast Fourier transform (FFT) analysis and one-third octave-band analysis were used for the detection of tonal low-frequency noise. Comparison between two different car makers was also done. Finally, the paper suggests some possibilities for scientifically assessing and evaluating low-frequency sounds in general, and some recommendations are introduced for scientific discussion, since sounds with strong low-frequency content (but not only strong) engender greater annoyance than is predicted by an A-weighted sound pressure level.

  17. Multiple sound source localization using gammatone auditory filtering and direct sound componence detection

    NASA Astrophysics Data System (ADS)

    Chen, Huaiyu; Cao, Li

    2017-06-01

    In order to research multiple sound source localization with room reverberation and background noise, we analyze the shortcomings of traditional broadband MUSIC and ordinary auditory filtering based broadband MUSIC method, then a new broadband MUSIC algorithm with gammatone auditory filtering of frequency component selection control and detection of ascending segment of direct sound componence is proposed. The proposed algorithm controls frequency component within the interested frequency band in multichannel bandpass filter stage. Detecting the direct sound componence of the sound source for suppressing room reverberation interference is also proposed, whose merits are fast calculation and avoiding using more complex de-reverberation processing algorithm. Besides, the pseudo-spectrum of different frequency channels is weighted by their maximum amplitude for every speech frame. Through the simulation and real room reverberation environment experiments, the proposed method has good performance. Dynamic multiple sound source localization experimental results indicate that the average absolute error of azimuth estimated by the proposed algorithm is less and the histogram result has higher angle resolution.

  18. Interior sound field control using generalized singular value decomposition in the frequency domain.

    PubMed

    Pasco, Yann; Gauthier, Philippe-Aubert; Berry, Alain; Moreau, Stéphane

    2017-01-01

    The problem of controlling a sound field inside a region surrounded by acoustic control sources is considered. Inspired by the Kirchhoff-Helmholtz integral, the use of double-layer source arrays allows such a control and avoids the modification of the external sound field by the control sources by the approximation of the sources as monopole and radial dipole transducers. However, the practical implementation of the Kirchhoff-Helmholtz integral in physical space leads to large numbers of control sources and error sensors along with excessive controller complexity in three dimensions. The present study investigates the potential of the Generalized Singular Value Decomposition (GSVD) to reduce the controller complexity and separate the effect of control sources on the interior and exterior sound fields, respectively. A proper truncation of the singular basis provided by the GSVD factorization is shown to lead to effective cancellation of the interior sound field at frequencies below the spatial Nyquist frequency of the control sources array while leaving the exterior sound field almost unchanged. Proofs of concept are provided through simulations achieved for interior problems by simulations in a free field scenario with circular arrays and in a reflective environment with square arrays.

  19. Battle Damage Modeling

    DTIC Science & Technology

    2010-05-01

    has been an increasing move towards armor systems which are both structural and protection components at the same time. Analysis of material response...the materials can move. As the FE analysis progresses the component will move while the mesh remains motionless (Figure 4). Individual nodes and cells...this parameter. This subroutine needs many inputs, such as the speed of sound in the material , the FE size mesh and the safety factor, which prevents

  20. Series expansions of rotating two and three dimensional sound fields.

    PubMed

    Poletti, M A

    2010-12-01

    The cylindrical and spherical harmonic expansions of oscillating sound fields rotating at a constant rate are derived. These expansions are a generalized form of the stationary sound field expansions. The derivations are based on the representation of interior and exterior sound fields using the simple source approach and determination of the simple source solutions with uniform rotation. Numerical simulations of rotating sound fields are presented to verify the theory.

  1. Approaches to the study of neural coding of sound source location and sound envelope in real environments

    PubMed Central

    Kuwada, Shigeyuki; Bishop, Brian; Kim, Duck O.

    2012-01-01

    The major functions of the auditory system are recognition (what is the sound) and localization (where is the sound). Although each of these has received considerable attention, rarely are they studied in combination. Furthermore, the stimuli used in the bulk of studies did not represent sound location in real environments and ignored the effects of reverberation. Another ignored dimension is the distance of a sound source. Finally, there is a scarcity of studies conducted in unanesthetized animals. We illustrate a set of efficient methods that overcome these shortcomings. We use the virtual auditory space method (VAS) to efficiently present sounds at different azimuths, different distances and in different environments. Additionally, this method allows for efficient switching between binaural and monaural stimulation and alteration of acoustic cues singly or in combination to elucidate neural mechanisms underlying localization and recognition. Such procedures cannot be performed with real sound field stimulation. Our research is designed to address the following questions: Are IC neurons specialized to process what and where auditory information? How does reverberation and distance of the sound source affect this processing? How do IC neurons represent sound source distance? Are neural mechanisms underlying envelope processing binaural or monaural? PMID:22754505

  2. Characterizing large river sounds: Providing context for understanding the environmental effects of noise produced by hydrokinetic turbines.

    PubMed

    Bevelhimer, Mark S; Deng, Z Daniel; Scherelis, Constantin

    2016-01-01

    Underwater noise associated with the installation and operation of hydrokinetic turbines in rivers and tidal zones presents a potential environmental concern for fish and marine mammals. Comparing the spectral quality of sounds emitted by hydrokinetic turbines to natural and other anthropogenic sound sources is an initial step at understanding potential environmental impacts. Underwater recordings were obtained from passing vessels and natural underwater sound sources in static and flowing waters. Static water measurements were taken in a lake with minimal background noise. Flowing water measurements were taken at a previously proposed deployment site for hydrokinetic turbines on the Mississippi River, where sounds created by flowing water are part of all measurements, both natural ambient and anthropogenic sources. Vessel sizes ranged from a small fishing boat with 60 hp outboard motor to an 18-unit barge train being pushed upstream by tugboat. As expected, large vessels with large engines created the highest sound levels, which were, on average, 40 dB greater than the sound created by an operating hydrokinetic turbine. A comparison of sound levels from the same sources at different distances using both spherical and cylindrical sound attenuation functions suggests that spherical model results more closely approximate observed sound attenuation.

  3. Sound Source Localization Using Non-Conformal Surface Sound Field Transformation Based on Spherical Harmonic Wave Decomposition

    PubMed Central

    Zhang, Lanyue; Ding, Dandan; Yang, Desen; Wang, Jia; Shi, Jie

    2017-01-01

    Spherical microphone arrays have been paid increasing attention for their ability to locate a sound source with arbitrary incident angle in three-dimensional space. Low-frequency sound sources are usually located by using spherical near-field acoustic holography. The reconstruction surface and holography surface are conformal surfaces in the conventional sound field transformation based on generalized Fourier transform. When the sound source is on the cylindrical surface, it is difficult to locate by using spherical surface conformal transform. The non-conformal sound field transformation by making a transfer matrix based on spherical harmonic wave decomposition is proposed in this paper, which can achieve the transformation of a spherical surface into a cylindrical surface by using spherical array data. The theoretical expressions of the proposed method are deduced, and the performance of the method is simulated. Moreover, the experiment of sound source localization by using a spherical array with randomly and uniformly distributed elements is carried out. Results show that the non-conformal surface sound field transformation from a spherical surface to a cylindrical surface is realized by using the proposed method. The localization deviation is around 0.01 m, and the resolution is around 0.3 m. The application of the spherical array is extended, and the localization ability of the spherical array is improved. PMID:28489065

  4. Sound reduction by metamaterial-based acoustic enclosure

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yao, Shanshan; Li, Pei; Zhou, Xiaoming

    In many practical systems, acoustic radiation control on noise sources contained within a finite volume by an acoustic enclosure is of great importance, but difficult to be accomplished at low frequencies due to the enhanced acoustic-structure interaction. In this work, we propose to use acoustic metamaterials as the enclosure to efficiently reduce sound radiation at their negative-mass frequencies. Based on a circularly-shaped metamaterial model, sound radiation properties by either central or eccentric sources are analyzed by numerical simulations for structured metamaterials. The parametric analyses demonstrate that the barrier thickness, the cavity size, the source type, and the eccentricity of themore » source have a profound effect on the sound reduction. It is found that increasing the thickness of the metamaterial barrier is an efficient approach to achieve large sound reduction over the negative-mass frequencies. These results are helpful in designing highly efficient acoustic enclosures for blockage of sound in low frequencies.« less

  5. Cross-correlation, triangulation, and curved-wavefront focusing of coral reef sound using a bi-linear hydrophone array.

    PubMed

    Freeman, Simon E; Buckingham, Michael J; Freeman, Lauren A; Lammers, Marc O; D'Spain, Gerald L

    2015-01-01

    A seven element, bi-linear hydrophone array was deployed over a coral reef in the Papahãnaumokuãkea Marine National Monument, Northwest Hawaiian Islands, in order to investigate the spatial, temporal, and spectral properties of biological sound in an environment free of anthropogenic influences. Local biological sound sources, including snapping shrimp and other organisms, produced curved-wavefront acoustic arrivals at the array, allowing source location via focusing to be performed over an area of 1600 m(2). Initially, however, a rough estimate of source location was obtained from triangulation of pair-wise cross-correlations of the sound. Refinements to these initial source locations, and source frequency information, were then obtained using two techniques, conventional and adaptive focusing. It was found that most of the sources were situated on or inside the reef structure itself, rather than over adjacent sandy areas. Snapping-shrimp-like sounds, all with similar spectral characteristics, originated from individual sources predominantly in one area to the east of the array. To the west, the spectral and spatial distributions of the sources were more varied, suggesting the presence of a multitude of heterogeneous biological processes. In addition to the biological sounds, some low-frequency noise due to distant breaking waves was received from end-fire north of the array.

  6. Underwater Acoustic Source Localisation Among Blind and Sighted Scuba Divers: Comparative study.

    PubMed

    Cambi, Jacopo; Livi, Ludovica; Livi, Walter

    2017-05-01

    Many blind individuals demonstrate enhanced auditory spatial discrimination or localisation of sound sources in comparison to sighted subjects. However, this hypothesis has not yet been confirmed with regards to underwater spatial localisation. This study therefore aimed to investigate underwater acoustic source localisation among blind and sighted scuba divers. This study took place between February and June 2015 in Elba, Italy, and involved two experimental groups of divers with either acquired (n = 20) or congenital (n = 10) blindness and a control group of 30 sighted divers. Each subject took part in five attempts at an under-water acoustic source localisation task, in which the divers were requested to swim to the source of a sound originating from one of 24 potential locations. The control group had their sight obscured during the task. The congenitally blind divers demonstrated significantly better underwater sound localisation compared to the control group or those with acquired blindness ( P = 0.0007). In addition, there was a significant correlation between years of blindness and underwater sound localisation ( P <0.0001). Congenital blindness was found to positively affect the ability of a diver to recognise the source of a sound in an underwater environment. As the correct localisation of sounds underwater may help individuals to avoid imminent danger, divers should perform sound localisation tests during training sessions.

  7. In situ analysis of measurements of auroral dynamics and structure

    NASA Astrophysics Data System (ADS)

    Mella, Meghan R.

    Two auroral sounding rocket case studies, one in the dayside and one in the nightside, explore aspects of poleward boundary aurora. The nightside sounding rocket, Cascades-2 was launched on 20 March 2009 at 11:04:00 UT from the Poker Flat Research Range in Alaska, and flew across a series of poleward boundary intensifications (PBIs). Each of the crossings have fundamentally different in situ electron energy and pitch angle structure, and different ground optics images of visible aurora. The different particle distributions show signatures of both a quasistatic acceleration mechanism and an Alfvenic acceleration mechanism, as well as combinations of both. The Cascades-2 experiment is the first sounding rocket observation of a PBI sequence, enabling a detailed investigation of the electron signatures and optical aurora associated with various stages of a PBI sequence as it evolves from an Alfvenic to a more quasistatic structure. The dayside sounding rocket, Scifer-2 was launched on 18 January 2008 at 7:30 UT from the Andoya Rocket Range in Andenes, Norway. It flew northward through the cleft region during a Poleward Moving Auroral Form (PMAF) event. Both the dayside and nightside flights observe dispersed, precipitating ions, each of a different nature. The dispersion signatures are dependent on, among other things, the MLT sector, altitude, source region, and precipitation mechanism. It is found that small changes in the shape of the dispersion have a large influence on whether the precipitation was localized or extended over a range of altitudes. It is also found that a single Maxwellian source will not replicate the data, but rather, a sum of Maxwellians of different temperature, similar to a Kappa distribution, most closely reproduces the data. The various particle signatures are used to argue that both events have similar magnetospheric drivers, that is, Bursty Bulk Flows in the magnetotail.

  8. "SMALLab": Virtual Geology Studies Using Embodied Learning with Motion, Sound, and Graphics

    ERIC Educational Resources Information Center

    Johnson-Glenberg, Mina C.; Birchfield, David; Usyal, Sibel

    2009-01-01

    We present a new and innovative interface that allows the learner's body to move freely in a multimodal learning environment. The Situated Multimedia Arts Learning Laboratory ("SMALLab") uses 3D object tracking, real time graphics, and surround-sound to enhance embodied learning. Our hypothesis is that optimal learning and retention occur when…

  9. 77 FR 49857 - Early Scoping Notification for the Alternatives Analysis of the Tacoma Link Expansion in Tacoma, WA

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-08-17

    ... requirements associated with the New Starts (``Section 5309'') funding program for certain kinds of major capital investments. While recent legislation may lead to changes in the New Starts process, Sound Transit... business and theater districts. Sound Move, the first phase of regional transit investments, was approved...

  10. The role of envelope shape in the localization of multiple sound sources and echoes in the barn owl.

    PubMed

    Baxter, Caitlin S; Nelson, Brian S; Takahashi, Terry T

    2013-02-01

    Echoes and sounds of independent origin often obscure sounds of interest, but echoes can go undetected under natural listening conditions, a perception called the precedence effect. How does the auditory system distinguish between echoes and independent sources? To investigate, we presented two broadband noises to barn owls (Tyto alba) while varying the similarity of the sounds' envelopes. The carriers of the noises were identical except for a 2- or 3-ms delay. Their onsets and offsets were also synchronized. In owls, sound localization is guided by neural activity on a topographic map of auditory space. When there are two sources concomitantly emitting sounds with overlapping amplitude spectra, space map neurons discharge when the stimulus in their receptive field is louder than the one outside it and when the averaged amplitudes of both sounds are rising. A model incorporating these features calculated the strengths of the two sources' representations on the map (B. S. Nelson and T. T. Takahashi; Neuron 67: 643-655, 2010). The target localized by the owls could be predicted from the model's output. The model also explained why the echo is not localized at short delays: when envelopes are similar, peaks in the leading sound mask corresponding peaks in the echo, weakening the echo's space map representation. When the envelopes are dissimilar, there are few or no corresponding peaks, and the owl localizes whichever source is predicted by the model to be less masked. Thus the precedence effect in the owl is a by-product of a mechanism for representing multiple sound sources on its map.

  11. Sensitivity to Angular and Radial Source Movements as a Function of Acoustic Complexity in Normal and Impaired Hearing

    PubMed Central

    Grimm, Giso; Hohmann, Volker; Laugesen, Søren; Neher, Tobias

    2017-01-01

    In contrast to static sounds, spatially dynamic sounds have received little attention in psychoacoustic research so far. This holds true especially for acoustically complex (reverberant, multisource) conditions and impaired hearing. The current study therefore investigated the influence of reverberation and the number of concurrent sound sources on source movement detection in young normal-hearing (YNH) and elderly hearing-impaired (EHI) listeners. A listening environment based on natural environmental sounds was simulated using virtual acoustics and rendered over headphones. Both near-far (‘radial’) and left-right (‘angular’) movements of a frontal target source were considered. The acoustic complexity was varied by adding static lateral distractor sound sources as well as reverberation. Acoustic analyses confirmed the expected changes in stimulus features that are thought to underlie radial and angular source movements under anechoic conditions and suggested a special role of monaural spectral changes under reverberant conditions. Analyses of the detection thresholds showed that, with the exception of the single-source scenarios, the EHI group was less sensitive to source movements than the YNH group, despite adequate stimulus audibility. Adding static sound sources clearly impaired the detectability of angular source movements for the EHI (but not the YNH) group. Reverberation, on the other hand, clearly impaired radial source movement detection for the EHI (but not the YNH) listeners. These results illustrate the feasibility of studying factors related to auditory movement perception with the help of the developed test setup. PMID:28675088

  12. Sensitivity to Angular and Radial Source Movements as a Function of Acoustic Complexity in Normal and Impaired Hearing.

    PubMed

    Lundbeck, Micha; Grimm, Giso; Hohmann, Volker; Laugesen, Søren; Neher, Tobias

    2017-01-01

    In contrast to static sounds, spatially dynamic sounds have received little attention in psychoacoustic research so far. This holds true especially for acoustically complex (reverberant, multisource) conditions and impaired hearing. The current study therefore investigated the influence of reverberation and the number of concurrent sound sources on source movement detection in young normal-hearing (YNH) and elderly hearing-impaired (EHI) listeners. A listening environment based on natural environmental sounds was simulated using virtual acoustics and rendered over headphones. Both near-far ('radial') and left-right ('angular') movements of a frontal target source were considered. The acoustic complexity was varied by adding static lateral distractor sound sources as well as reverberation. Acoustic analyses confirmed the expected changes in stimulus features that are thought to underlie radial and angular source movements under anechoic conditions and suggested a special role of monaural spectral changes under reverberant conditions. Analyses of the detection thresholds showed that, with the exception of the single-source scenarios, the EHI group was less sensitive to source movements than the YNH group, despite adequate stimulus audibility. Adding static sound sources clearly impaired the detectability of angular source movements for the EHI (but not the YNH) group. Reverberation, on the other hand, clearly impaired radial source movement detection for the EHI (but not the YNH) listeners. These results illustrate the feasibility of studying factors related to auditory movement perception with the help of the developed test setup.

  13. A Computational and Experimental Study of Resonators in Three Dimensions

    NASA Technical Reports Server (NTRS)

    Tam, C. K. W.; Ju, H.; Jones, Michael G.; Watson, Willie R.; Parrott, Tony L.

    2009-01-01

    In a previous work by the present authors, a computational and experimental investigation of the acoustic properties of two-dimensional slit resonators was carried out. The present paper reports the results of a study extending the previous work to three dimensions. This investigation has two basic objectives. The first is to validate the computed results from direct numerical simulations of the flow and acoustic fields of slit resonators in three dimensions by comparing with experimental measurements in a normal incidence impedance tube. The second objective is to study the flow physics of resonant liners responsible for sound wave dissipation. Extensive comparisons are provided between computed and measured acoustic liner properties with both discrete frequency and broadband sound sources. Good agreements are found over a wide range of frequencies and sound pressure levels. Direct numerical simulation confirms the previous finding in two dimensions that vortex shedding is the dominant dissipation mechanism at high sound pressure intensity. However, it is observed that the behavior of the shed vortices in three dimensions is quite different from those of two dimensions. In three dimensions, the shed vortices tend to evolve into ring (circular in plan form) vortices, even though the slit resonator opening from which the vortices are shed has an aspect ratio of 2.5. Under the excitation of discrete frequency sound, the shed vortices align themselves into two regularly spaced vortex trains moving away from the resonator opening in opposite directions. This is different from the chaotic shedding of vortices found in two-dimensional simulations. The effect of slit aspect ratio at a fixed porosity is briefly studied. For the range of liners considered in this investigation, it is found that the absorption coefficient of a liner increases when the open area of the single slit is subdivided into multiple, smaller slits.

  14. Vestibular Influence on Auditory Metrical Interpretation

    ERIC Educational Resources Information Center

    Phillips-Silver, J.; Trainor, L.J.

    2008-01-01

    When we move to music we feel the beat, and this feeling can shape the sound we hear. Previous studies have shown that when people listen to a metrically ambiguous rhythm pattern, moving the body on a certain beat-adults, by actively bouncing themselves in synchrony with the experimenter, and babies, by being bounced passively in the…

  15. 30 CFR 56.14200 - Warnings prior to starting or moving equipment.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 30 Mineral Resources 1 2011-07-01 2011-07-01 false Warnings prior to starting or moving equipment... MINES Machinery and Equipment Safety Practices and Operational Procedures § 56.14200 Warnings prior to... operators shall sound a warning that is audible above the surrounding noise level or use other effective...

  16. 30 CFR 56.14200 - Warnings prior to starting or moving equipment.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 30 Mineral Resources 1 2014-07-01 2014-07-01 false Warnings prior to starting or moving equipment... MINES Machinery and Equipment Safety Practices and Operational Procedures § 56.14200 Warnings prior to... operators shall sound a warning that is audible above the surrounding noise level or use other effective...

  17. 30 CFR 56.14200 - Warnings prior to starting or moving equipment.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 30 Mineral Resources 1 2013-07-01 2013-07-01 false Warnings prior to starting or moving equipment... MINES Machinery and Equipment Safety Practices and Operational Procedures § 56.14200 Warnings prior to... operators shall sound a warning that is audible above the surrounding noise level or use other effective...

  18. 30 CFR 56.14200 - Warnings prior to starting or moving equipment.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Warnings prior to starting or moving equipment... MINES Machinery and Equipment Safety Practices and Operational Procedures § 56.14200 Warnings prior to... operators shall sound a warning that is audible above the surrounding noise level or use other effective...

  19. Apparatus for the remote detection of sounds caused by leaks

    NASA Technical Reports Server (NTRS)

    Shakkottai, Parthasarathy (Inventor); Alwar, Vijayaraghavan (Inventor)

    1990-01-01

    Two laser beams derived from a laser pass through a furnace and are superposed in a laser spot to produce interference fringes having the same pitch as that of a Ronchi grating printed on a retroreflecting screen. Minute fluctuations of the laser beams caused by sound waves from leaks result in intensity fluctuations of the laser spot when the laser fringes move past the fixed grating. A telescope and photocell arrangement detects the light variations to produce an electrical signal which is amplified and filtered to detect the sound of leaks. This non-contact laser Schlieren microphone is sensitive to sounds all along the path of the beams.

  20. Ultrasonic Recovery and Modification of Food Ingredients

    NASA Astrophysics Data System (ADS)

    Vilkhu, Kamaljit; Manasseh, Richard; Mawson, Raymond; Ashokkumar, Muthupandian

    There are two general classes of effects that sound, and ultrasound in particular, can have on a fluid. First, very significant modifications to the nature of food and food ingredients can be due to the phenomena of bubble acoustics and cavitation. The applied sound oscillates bubbles in the fluid, creating intense forces at microscopic scales thus driving chemical changes. Second, the sound itself can cause the fluid to flow vigorously, both on a large scale and on a microscopic scale; furthermore, the sound can cause particles in the fluid to move relative to the fluid. These streaming phenomena can redistribute materials within food and food ingredients at both microscopic and macroscopic scales.

  1. Localization and tracking of moving objects in two-dimensional space by echolocation.

    PubMed

    Matsuo, Ikuo

    2013-02-01

    Bats use frequency-modulated echolocation to identify and capture moving objects in real three-dimensional space. Experimental evidence indicates that bats are capable of locating static objects with a range accuracy of less than 1 μs. A previously introduced model estimates ranges of multiple, static objects using linear frequency modulation (LFM) sound and Gaussian chirplets with a carrier frequency compatible with bat emission sweep rates. The delay time for a single object was estimated with an accuracy of about 1.3 μs by measuring the echo at a low signal-to-noise ratio (SNR). The range accuracy was dependent not only on the SNR but also the Doppler shift, which was dependent on the movements. However, it was unclear whether this model could estimate the moving object range at each timepoint. In this study, echoes were measured from the rotating pole at two receiving points by intermittently emitting LFM sounds. The model was shown to localize moving objects in two-dimensional space by accurately estimating the object's range at each timepoint.

  2. Characterizing large river sounds: Providing context for understanding the environmental effects of noise produced by hydrokinetic turbines

    DOE PAGES

    Bevelhimer, Mark S.; Deng, Z. Daniel; Scherelis, Constantin C.

    2016-01-06

    Underwaternoise associated with the installation and operation of hydrokinetic turbines in rivers and tidal zones presents a potential environmental concern for fish and marine mammals. Comparing the spectral quality of sounds emitted by hydrokinetic turbines to natural and other anthropogenic sound sources is an initial step at understanding potential environmental impacts. Underwater recordings were obtained from passing vessels and natural underwater sound sources in static and flowing waters. Static water measurements were taken in a lake with minimal background noise. Flowing water measurements were taken at a previously proposed deployment site for hydrokinetic turbines on the Mississippi River, where soundsmore » created by flowing water are part of all measurements, both natural ambient and anthropogenic sources. Vessel sizes ranged from a small fishing boat with 60 hp outboard motor to an 18-unit barge train being pushed upstream by tugboat. As expected, large vessels with large engines created the highest sound levels, which were, on average, 40 dB greater than the sound created by an operating hydrokinetic turbine. As a result, a comparison of sound levels from the same sources at different distances using both spherical and cylindrical sound attenuation functions suggests that spherical model results more closely approximate observed sound attenuation.« less

  3. Underwater auditory localization by a swimming harbor seal (Phoca vitulina).

    PubMed

    Bodson, Anais; Miersch, Lars; Mauck, Bjoern; Dehnhardt, Guido

    2006-09-01

    The underwater sound localization acuity of a swimming harbor seal (Phoca vitulina) was measured in the horizontal plane at 13 different positions. The stimulus was either a double sound (two 6-kHz pure tones lasting 0.5 s separated by an interval of 0.2 s) or a single continuous sound of 1.2 s. Testing was conducted in a 10-m-diam underwater half circle arena with hidden loudspeakers installed at the exterior perimeter. The animal was trained to swim along the diameter of the half circle and to change its course towards the sound source as soon as the signal was given. The seal indicated the sound source by touching its assumed position at the board of the half circle. The deviation of the seals choice from the actual sound source was measured by means of video analysis. In trials with the double sound the seal localized the sound sources with a mean deviation of 2.8 degrees and in trials with the single sound with a mean deviation of 4.5 degrees. In a second experiment minimum audible angles of the stationary animal were found to be 9.8 degrees in front and 9.7 degrees in the back of the seal's head.

  4. Personal sound zone reproduction with room reflections

    NASA Astrophysics Data System (ADS)

    Olik, Marek

    Loudspeaker-based sound systems, capable of a convincing reproduction of different audio streams to listeners in the same acoustic enclosure, are a convenient alternative to headphones. Such systems aim to generate "sound zones" in which target sound programmes are to be reproduced with minimum interference from any alternative programmes. This can be achieved with appropriate filtering of the source (loudspeaker) signals, so that the target sound's energy is directed to the chosen zone while being attenuated elsewhere. The existing methods are unable to produce the required sound energy ratio (acoustic contrast) between the zones with a small number of sources when strong room reflections are present. Optimization of parameters is therefore required for systems with practical limitations to improve their performance in reflective acoustic environments. One important parameter is positioning of sources with respect to the zones and room boundaries. The first contribution of this thesis is a comparison of the key sound zoning methods implemented on compact and distributed geometrical source arrangements. The study presents previously unpublished detailed evaluation and ranking of such arrangements for systems with a limited number of sources in a reflective acoustic environment similar to a domestic room. Motivated by the requirement to investigate the relationship between source positioning and performance in detail, the central contribution of this thesis is a study on optimizing source arrangements when strong individual room reflections occur. Small sound zone systems are studied analytically and numerically to reveal relationships between the geometry of source arrays and performance in terms of acoustic contrast and array effort (related to system efficiency). Three novel source position optimization techniques are proposed to increase the contrast, and geometrical means of reducing the effort are determined. Contrary to previously published case studies, this work presents a systematic examination of the key problem of first order reflections and proposes general optimization techniques, thus forming an important contribution. The remaining contribution considers evaluation and comparison of the proposed techniques with two alternative approaches to sound zone generation under reflective conditions: acoustic contrast control (ACC) combined with anechoic source optimization and sound power minimization (SPM). The study provides a ranking of the examined approaches which could serve as a guideline for method selection for rooms with strong individual reflections.

  5. Marine mammal audibility of selected shallow-water survey sources.

    PubMed

    MacGillivray, Alexander O; Racca, Roberto; Li, Zizheng

    2014-01-01

    Most attention about the acoustic effects of marine survey sound sources on marine mammals has focused on airgun arrays, with other common sources receiving less scrutiny. Sound levels above hearing threshold (sensation levels) were modeled for six marine mammal species and seven different survey sources in shallow water. The model indicated that odontocetes were most likely to hear sounds from mid-frequency sources (fishery, communication, and hydrographic systems), mysticetes from low-frequency sources (sub-bottom profiler and airguns), and pinnipeds from both mid- and low-frequency sources. High-frequency sources (side-scan and multibeam) generated the lowest estimated sensation levels for all marine mammal species groups.

  6. Performance of active feedforward control systems in non-ideal, synthesized diffuse sound fields.

    PubMed

    Misol, Malte; Bloch, Christian; Monner, Hans Peter; Sinapius, Michael

    2014-04-01

    The acoustic performance of passive or active panel structures is usually tested in sound transmission loss facilities. A reverberant sending room, equipped with one or a number of independent sound sources, is used to generate a diffuse sound field excitation which acts as a disturbance source on the structure under investigation. The spatial correlation and coherence of such a synthesized non-ideal diffuse-sound-field excitation, however, might deviate significantly from the ideal case. This has consequences for the operation of an active feedforward control system which heavily relies on the acquisition of coherent disturbance source information. This work, therefore, evaluates the spatial correlation and coherence of ideal and non-ideal diffuse sound fields and considers the implications on the performance of a feedforward control system. The system under consideration is an aircraft-typical double panel system, equipped with an active sidewall panel (lining), which is realized in a transmission loss facility. Experimental results for different numbers of sound sources in the reverberation room are compared to simulation results of a comparable generic double panel system excited by an ideal diffuse sound field. It is shown that the number of statistically independent noise sources acting on the primary structure of the double panel system depends not only on the type of diffuse sound field but also on the sample lengths of the processed signals. The experimental results show that the number of reference sensors required for a defined control performance exhibits an inverse relationship to control filter length.

  7. The influence of underwater data transmission sounds on the displacement behaviour of captive harbour seals (Phoca vitulina).

    PubMed

    Kastelein, Ronald A; van der Heul, Sander; Verboom, Willem C; Triesscheijn, Rob J V; Jennings, Nancy V

    2006-02-01

    To prevent grounding of ships and collisions between ships in shallow coastal waters, an underwater data collection and communication network (ACME) using underwater sounds to encode and transmit data is currently under development. Marine mammals might be affected by ACME sounds since they may use sound of a similar frequency (around 12 kHz) for communication, orientation, and prey location. If marine mammals tend to avoid the vicinity of the acoustic transmitters, they may be kept away from ecologically important areas by ACME sounds. One marine mammal species that may be affected in the North Sea is the harbour seal (Phoca vitulina). No information is available on the effects of ACME-like sounds on harbour seals, so this study was carried out as part of an environmental impact assessment program. Nine captive harbour seals were subjected to four sound types, three of which may be used in the underwater acoustic data communication network. The effect of each sound was judged by comparing the animals' location in a pool during test periods to that during baseline periods, during which no sound was produced. Each of the four sounds could be made into a deterrent by increasing its amplitude. The seals reacted by swimming away from the sound source. The sound pressure level (SPL) at the acoustic discomfort threshold was established for each of the four sounds. The acoustic discomfort threshold is defined as the boundary between the areas that the animals generally occupied during the transmission of the sounds and the areas that they generally did not enter during transmission. The SPLs at the acoustic discomfort thresholds were similar for each of the sounds (107 dB re 1 microPa). Based on this discomfort threshold SPL, discomfort zones at sea for several source levels (130-180 dB re 1 microPa) of the sounds were calculated, using a guideline sound propagation model for shallow water. The discomfort zone is defined as the area around a sound source that harbour seals are expected to avoid. The definition of the discomfort zone is based on behavioural discomfort, and does not necessarily coincide with the physical discomfort zone. Based on these results, source levels can be selected that have an acceptable effect on harbour seals in particular areas. The discomfort zone of a communication sound depends on the sound, the source level, and the propagation characteristics of the area in which the sound system is operational. The source level of the communication system should be adapted to each area (taking into account the width of a sea arm, the local sound propagation, and the importance of an area to the affected species). The discomfort zone should not coincide with ecologically important areas (for instance resting, breeding, suckling, and feeding areas), or routes between these areas.

  8. Monitoring Anthropogenic Ocean Sound from Shipping Using an Acoustic Sensor Network and a Compressive Sensing Approach.

    PubMed

    Harris, Peter; Philip, Rachel; Robinson, Stephen; Wang, Lian

    2016-03-22

    Monitoring ocean acoustic noise has been the subject of considerable recent study, motivated by the desire to assess the impact of anthropogenic noise on marine life. A combination of measuring ocean sound using an acoustic sensor network and modelling sources of sound and sound propagation has been proposed as an approach to estimating the acoustic noise map within a region of interest. However, strategies for developing a monitoring network are not well established. In this paper, considerations for designing a network are investigated using a simulated scenario based on the measurement of sound from ships in a shipping lane. Using models for the sources of the sound and for sound propagation, a noise map is calculated and measurements of the noise map by a sensor network within the region of interest are simulated. A compressive sensing algorithm, which exploits the sparsity of the representation of the noise map in terms of the sources, is used to estimate the locations and levels of the sources and thence the entire noise map within the region of interest. It is shown that although the spatial resolution to which the sound sources can be identified is generally limited, estimates of aggregated measures of the noise map can be obtained that are more reliable compared with those provided by other approaches.

  9. Monitoring Anthropogenic Ocean Sound from Shipping Using an Acoustic Sensor Network and a Compressive Sensing Approach †

    PubMed Central

    Harris, Peter; Philip, Rachel; Robinson, Stephen; Wang, Lian

    2016-01-01

    Monitoring ocean acoustic noise has been the subject of considerable recent study, motivated by the desire to assess the impact of anthropogenic noise on marine life. A combination of measuring ocean sound using an acoustic sensor network and modelling sources of sound and sound propagation has been proposed as an approach to estimating the acoustic noise map within a region of interest. However, strategies for developing a monitoring network are not well established. In this paper, considerations for designing a network are investigated using a simulated scenario based on the measurement of sound from ships in a shipping lane. Using models for the sources of the sound and for sound propagation, a noise map is calculated and measurements of the noise map by a sensor network within the region of interest are simulated. A compressive sensing algorithm, which exploits the sparsity of the representation of the noise map in terms of the sources, is used to estimate the locations and levels of the sources and thence the entire noise map within the region of interest. It is shown that although the spatial resolution to which the sound sources can be identified is generally limited, estimates of aggregated measures of the noise map can be obtained that are more reliable compared with those provided by other approaches. PMID:27011187

  10. Viscous Torques on a Levitating Body

    NASA Technical Reports Server (NTRS)

    Busse, F.; Wang, T.

    1982-01-01

    New analytical expressions for viscous torque generated by orthogonal sound waves agree well with experiment. It is possible to calculate torque on an object levitated in a fluid. Levitation has applications in containerless materials processing, coating, and fabrication of small precision parts. Sound waves cause fluid particles to move in elliptical paths and induce azimuthal circulation in boundary layer, giving rise to time-averaged torque.

  11. Underwater Acoustic Source Localisation Among Blind and Sighted Scuba Divers

    PubMed Central

    Cambi, Jacopo; Livi, Ludovica; Livi, Walter

    2017-01-01

    Objectives Many blind individuals demonstrate enhanced auditory spatial discrimination or localisation of sound sources in comparison to sighted subjects. However, this hypothesis has not yet been confirmed with regards to underwater spatial localisation. This study therefore aimed to investigate underwater acoustic source localisation among blind and sighted scuba divers. Methods This study took place between February and June 2015 in Elba, Italy, and involved two experimental groups of divers with either acquired (n = 20) or congenital (n = 10) blindness and a control group of 30 sighted divers. Each subject took part in five attempts at an under-water acoustic source localisation task, in which the divers were requested to swim to the source of a sound originating from one of 24 potential locations. The control group had their sight obscured during the task. Results The congenitally blind divers demonstrated significantly better underwater sound localisation compared to the control group or those with acquired blindness (P = 0.0007). In addition, there was a significant correlation between years of blindness and underwater sound localisation (P <0.0001). Conclusion Congenital blindness was found to positively affect the ability of a diver to recognise the source of a sound in an underwater environment. As the correct localisation of sounds underwater may help individuals to avoid imminent danger, divers should perform sound localisation tests during training sessions. PMID:28690888

  12. A general introduction to aeroacoustics and atmospheric sound

    NASA Technical Reports Server (NTRS)

    Lighthill, James

    1992-01-01

    A single unifying principle (based upon the nonlinear 'momentum-flux' effects produced when different components of a motion transport different components of its momentum) is used to give a broad scientific background to several aspects of the interaction between airflows and atmospheric sound. First, it treats the generation of sound by airflows of many different types. These include, for example, jet-like flows involving convected turbulent motions (with the resulting aeroacoustic radiation sensitively dependent on the Mach number of convection) and they include, as an extreme case, the supersonic 'boom' (shock waves generated by a supersonically convected flow pattern). Next, an analysis is given of sound propagation through nonuniformly moving airflows, and the exchange is quantified of energy between flow and sound; while, finally, problems are examined of how sound waves 'on their own' may generate the airflows known as acoustic streaming.

  13. The extended Fourier pseudospectral time-domain method for atmospheric sound propagation.

    PubMed

    Hornikx, Maarten; Waxler, Roger; Forssén, Jens

    2010-10-01

    An extended Fourier pseudospectral time-domain (PSTD) method is presented to model atmospheric sound propagation by solving the linearized Euler equations. In this method, evaluation of spatial derivatives is based on an eigenfunction expansion. Evaluation on a spatial grid requires only two spatial points per wavelength. Time iteration is done using a low-storage optimized six-stage Runge-Kutta method. This method is applied to two-dimensional non-moving media models, one with screens and one for an urban canyon, with generally high accuracy in both amplitude and phase. For a moving atmosphere, accurate results have been obtained in models with both a uniform and a logarithmic wind velocity profile over a rigid ground surface and in the presence of a screen. The method has also been validated for three-dimensional sound propagation over a screen. For that application, the developed method is in the order of 100 times faster than the second-order-accurate FDTD solution to the linearized Euler equations. The method is found to be well suited for atmospheric sound propagation simulations where effects of complex meteorology and straight rigid boundary surfaces are to be investigated.

  14. The role of envelope shape in the localization of multiple sound sources and echoes in the barn owl

    PubMed Central

    Baxter, Caitlin S.; Takahashi, Terry T.

    2013-01-01

    Echoes and sounds of independent origin often obscure sounds of interest, but echoes can go undetected under natural listening conditions, a perception called the precedence effect. How does the auditory system distinguish between echoes and independent sources? To investigate, we presented two broadband noises to barn owls (Tyto alba) while varying the similarity of the sounds' envelopes. The carriers of the noises were identical except for a 2- or 3-ms delay. Their onsets and offsets were also synchronized. In owls, sound localization is guided by neural activity on a topographic map of auditory space. When there are two sources concomitantly emitting sounds with overlapping amplitude spectra, space map neurons discharge when the stimulus in their receptive field is louder than the one outside it and when the averaged amplitudes of both sounds are rising. A model incorporating these features calculated the strengths of the two sources' representations on the map (B. S. Nelson and T. T. Takahashi; Neuron 67: 643–655, 2010). The target localized by the owls could be predicted from the model's output. The model also explained why the echo is not localized at short delays: when envelopes are similar, peaks in the leading sound mask corresponding peaks in the echo, weakening the echo's space map representation. When the envelopes are dissimilar, there are few or no corresponding peaks, and the owl localizes whichever source is predicted by the model to be less masked. Thus the precedence effect in the owl is a by-product of a mechanism for representing multiple sound sources on its map. PMID:23175801

  15. Comparison of sound reproduction using higher order loudspeakers and equivalent line arrays in free-field conditions.

    PubMed

    Poletti, Mark A; Betlehem, Terence; Abhayapala, Thushara D

    2014-07-01

    Higher order sound sources of Nth order can radiate sound with 2N + 1 orthogonal radiation patterns, which can be represented as phase modes or, equivalently, amplitude modes. This paper shows that each phase mode response produces a spiral wave front with a different spiral rate, and therefore a different direction of arrival of sound. Hence, for a given receiver position a higher order source is equivalent to a linear array of 2N + 1 monopole sources. This interpretation suggests performance similar to a circular array of higher order sources can be produced by an array of sources, each of which consists of a line array having monopoles at the apparent source locations of the corresponding phase modes. Simulations of higher order arrays and arrays of equivalent line sources are presented. It is shown that the interior fields produced by the two arrays are essentially the same, but that the exterior fields differ because the higher order sources produces different equivalent source locations for field positions outside the array. This work provides an explanation of the fact that an array of L Nth order sources can reproduce sound fields whose accuracy approaches the performance of (2N + 1)L monopoles.

  16. Structure of supersonic jet flow and its radiated sound

    NASA Technical Reports Server (NTRS)

    Mankbadi, Reda R.; Hayer, M. Ehtesham; Povinelli, Louis A.

    1994-01-01

    The present paper explores the use of large-eddy simulations as a tool for predicting noise from first principles. A high-order numerical scheme is used to perform large-eddy simulations of a supersonic jet flow with emphasis on capturing the time-dependent flow structure representating the sound source. The wavelike nature of this structure under random inflow disturbances is demonstrated. This wavelike structure is then enhanced by taking the inflow disturbances to be purely harmonic. Application of Lighthill's theory to calculate the far-field noise, with the sound source obtained from the calculated time-dependent near field, is demonstrated. Alternative approaches to coupling the near-field sound source to the far-field sound are discussed.

  17. Spacecraft Internal Acoustic Environment Modeling

    NASA Technical Reports Server (NTRS)

    Chu, SShao-sheng R.; Allen, Christopher S.

    2009-01-01

    Acoustic modeling can be used to identify key noise sources, determine/analyze sub-allocated requirements, keep track of the accumulation of minor noise sources, and to predict vehicle noise levels at various stages in vehicle development, first with estimates of noise sources, later with experimental data. In FY09, the physical mockup developed in FY08, with interior geometric shape similar to Orion CM (Crew Module) IML (Interior Mode Line), was used to validate SEA (Statistical Energy Analysis) acoustic model development with realistic ventilation fan sources. The sound power levels of these sources were unknown a priori, as opposed to previous studies that RSS (Reference Sound Source) with known sound power level was used. The modeling results were evaluated based on comparisons to measurements of sound pressure levels over a wide frequency range, including the frequency range where SEA gives good results. Sound intensity measurement was performed over a rectangular-shaped grid system enclosing the ventilation fan source. Sound intensities were measured at the top, front, back, right, and left surfaces of the and system. Sound intensity at the bottom surface was not measured, but sound blocking material was placed tinder the bottom surface to reflect most of the incident sound energy back to the remaining measured surfaces. Integrating measured sound intensities over measured surfaces renders estimated sound power of the source. The reverberation time T6o of the mockup interior had been modified to match reverberation levels of ISS US Lab interior for speech frequency bands, i.e., 0.5k, 1k, 2k, 4 kHz, by attaching appropriately sized Thinsulate sound absorption material to the interior wall of the mockup. Sound absorption of Thinsulate was modeled in three methods: Sabine equation with measured mockup interior reverberation time T60, layup model based on past impedance tube testing, and layup model plus air absorption correction. The evaluation/validation was carried out by acquiring octave band microphone data simultaneously at ten fixed locations throughout the mockup. SPLs (Sound Pressure Levels) predicted by our SEA model match well with measurements for our CM mockup, with a more complicated shape. Additionally in FY09, background NC noise (Noise Criterion) simulation and MRT (Modified Rhyme Test) were developed and performed in the mockup to determine the maximum noise level in CM habitable volume for fair crew voice communications. Numerous demonstrations of simulated noise environment in the mockup and associated SIL (Speech Interference Level) via MRT were performed for various communities, including members from NASA and Orion prime-/sub-contractors. Also, a new HSIR (Human-Systems Integration Requirement) for limiting pre- and post-landing SIL was proposed.

  18. A more accurate method using MOVES (Motor Vehicle Emission Simulator) to estimate emission burden for regional-level analysis.

    PubMed

    Liu, Xiaobo

    2015-07-01

    The U.S. Environmental Protection Agency's (EPA) Motor Vehicle Emission Simulator (MOVES) is required by the EPA to replace Mobile 6 as an official on-road emission model. Incorporated with annual vehicle mile traveled (VMT) by Highways Performance Monitoring System (HPMS) vehicle class, MOVES allocates VMT from HPMS to MOVES source (vehicle) types and calculates emission burden by MOVES source type. However, the calculated running emission burden by MOVES source type may be deviated from the actual emission burden because of MOVES source population, specifically the population fraction by MOVES source type in HPMS vehicle class. The deviation is also the result of the use of the universal set of parameters, i.e., relative mileage accumulation rate (relativeMAR), packaged in MOVES default database. This paper presents a novel approach by adjusting the relativeMAR to eliminate the impact of MOVES source population on running exhaust emission and to keep start and evaporative emissions unchanged for both MOVES2010b and MOVES2014. Results from MOVES runs using this approach indicated significant improvements on VMT distribution and emission burden estimation for each MOVES source type. The deviation of VMT by MOVES source type is minimized by using this approach from 12% to less than 0.05% for MOVES2010b and from 50% to less than 0.2% for MOVES2014 except for MOVES source type 53. Source type 53 still remains about 30% variation. The improvement of VMT distribution results in the elimination of emission burden deviation for each MOVES source type. For MOVES2010b, the deviation of emission burdens decreases from -12% for particulate matter less than 2.5 μm (PM2.5) and -9% for carbon monoxide (CO) to less than 0.002%. For MOVES2014, it drops from 80% for CO and 97% for PM2.5 to 0.006%. This approach is developed to more accurately estimate the total emission burdens using EPA's MOVES, both MOVES2010b and MOVES2014, by redistributing vehicle mile traveled (VMT) by Highways Performance Monitoring System (HPMS) class to MOVES source type on the basis of comprehensive traffic study, local link-by-link VMT broken down into MOVES source type.

  19. Underwater sound of rigid-hulled inflatable boats.

    PubMed

    Erbe, Christine; Liong, Syafrin; Koessler, Matthew Walter; Duncan, Alec J; Gourlay, Tim

    2016-06-01

    Underwater sound of rigid-hulled inflatable boats was recorded 142 times in total, over 3 sites: 2 in southern British Columbia, Canada, and 1 off Western Australia. Underwater sound peaked between 70 and 400 Hz, exhibiting strong tones in this frequency range related to engine and propeller rotation. Sound propagation models were applied to compute monopole source levels, with the source assumed 1 m below the sea surface. Broadband source levels (10-48 000 Hz) increased from 134 to 171 dB re 1 μPa @ 1 m with speed from 3 to 16 m/s (10-56 km/h). Source power spectral density percentile levels and 1/3 octave band levels are given for use in predictive modeling of underwater sound of these boats as part of environmental impact assessments.

  20. Sounds produced by individual white whales, Delphinapterus leucas, from Svalbard during capture (L)

    NASA Astrophysics Data System (ADS)

    van Parijs, Sofie M.; Lydersen, Christian; Kovacs, Kit M.

    2003-01-01

    Recordings were made of the sounds produced by white whales during capture events in Storfjorden, Svalbard, in the late autumn. Only four of eight captured individuals produced sounds. Four subadults, one female and three males, between 330 and 375 cm long, did not produce sounds during handling. The four animals that produced sounds were as follows: a female subadult of 280 cm produced repetitive broadband clicks; a solitary calf produced harmonic sounds, which we suggest may serve as mother-calf ``contact calls,'' and a mother-calf pair were the two animals that produced the most sounds in the study. The mother produced ``crooning'' broadband clicks and frequently moved her head toward her calf while producing underwater sounds. The calf produced three types of frequency-modulated sounds interspersed within broadband click trains. No sounds were heard from any of the animals once they were free-swimming, or during ad lib recording sessions in the study area, even though groups of white whales were sighted on several occasions away from the capture net.

  1. Control of Toxic Chemicals in Puget Sound, Phase 3: Study Of Atmospheric Deposition of Air Toxics to the Surface of Puget Sound

    DTIC Science & Technology

    2007-01-01

    deposition directly to Puget Sound was an important source of PAHs, polybrominated diphenyl ethers (PBDEs), and heavy metals . In most cases, atmospheric...versus Atmospheric Fluxes ........................................................................66  PAH Source Apportionment ...temperature inversions) on air quality during the wet season. A semi-quantitative apportionment study permitted a first-order characterization of source

  2. Psychophysical investigation of an auditory spatial illusion in cats: the precedence effect.

    PubMed

    Tollin, Daniel J; Yin, Tom C T

    2003-10-01

    The precedence effect (PE) describes several spatial perceptual phenomena that occur when similar sounds are presented from two different locations and separated by a delay. The mechanisms that produce the effect are thought to be responsible for the ability to localize sounds in reverberant environments. Although the physiological bases for the PE have been studied, little is known about how these sounds are localized by species other than humans. Here we used the search coil technique to measure the eye positions of cats trained to saccade to the apparent locations of sounds. To study the PE, brief broadband stimuli were presented from two locations, with a delay between their onsets; the delayed sound meant to simulate a single reflection. Although the cats accurately localized single sources, the apparent locations of the paired sources depended on the delay. First, the cats exhibited summing localization, the perception of a "phantom" sound located between the sources, for delays < +/-400 micros for sources positioned in azimuth along the horizontal plane, but not for sources positioned in elevation along the sagittal plane. Second, consistent with localization dominance, for delays from 400 micros to about 10 ms, the cats oriented toward the leading source location only, with little influence of the lagging source, both for horizontally and vertically placed sources. Finally, the echo threshold was reached for delays >10 ms, where the cats first began to orient to the lagging source on some trials. These data reveal that cats experience the PE phenomena similarly to humans.

  3. The Specificity of Sound Symbolic Correspondences in Spoken Language.

    PubMed

    Tzeng, Christina Y; Nygaard, Lynne C; Namy, Laura L

    2017-11-01

    Although language has long been regarded as a primarily arbitrary system, sound symbolism, or non-arbitrary correspondences between the sound of a word and its meaning, also exists in natural language. Previous research suggests that listeners are sensitive to sound symbolism. However, little is known about the specificity of these mappings. This study investigated whether sound symbolic properties correspond to specific meanings, or whether these properties generalize across semantic dimensions. In three experiments, native English-speaking adults heard sound symbolic foreign words for dimensional adjective pairs (big/small, round/pointy, fast/slow, moving/still) and for each foreign word, selected a translation among English antonyms that either matched or mismatched with the correct meaning dimension. Listeners agreed more reliably on the English translation for matched relative to mismatched dimensions, though reliable cross-dimensional mappings did occur. These findings suggest that although sound symbolic properties generalize to meanings that may share overlapping semantic features, sound symbolic mappings offer semantic specificity. Copyright © 2016 Cognitive Science Society, Inc.

  4. Visualizing Sound: Demonstrations to Teach Acoustic Concepts

    NASA Astrophysics Data System (ADS)

    Rennoll, Valerie

    Interference, a phenomenon in which two sound waves superpose to form a resultant wave of greater or lower amplitude, is a key concept when learning about the physics of sound waves. Typical interference demonstrations involve students listening for changes in sound level as they move throughout a room. Here, new tools are developed to teach this concept that provide a visual component, allowing individuals to see changes in sound level on a light display. This is accomplished using a microcontroller that analyzes sound levels collected by a microphone and displays the sound level in real-time on an LED strip. The light display is placed on a sliding rail between two speakers to show the interference occurring between two sound waves. When a long-exposure photograph is taken of the light display being slid from one end of the rail to the other, a wave of the interference pattern can be captured. By providing a visual component, these tools will help students and the general public to better understand interference, a key concept in acoustics.

  5. Evidence of sound production by spawning lake trout (Salvelinus namaycush) in lakes Huron and Champlain

    USGS Publications Warehouse

    Johnson, Nicholas S.; Higgs, Dennis; Binder, Thomas R.; Marsden, J. Ellen; Buchinger, Tyler John; Brege, Linnea; Bruning, Tyler; Farha, Steve A.; Krueger, Charles C.

    2018-01-01

    Two sounds associated with spawning lake trout (Salvelinus namaycush) in lakes Huron and Champlain were characterized by comparing sound recordings to behavioral data collected using acoustic telemetry and video. These sounds were named growls and snaps, and were heard on lake trout spawning reefs, but not on a non-spawning reef, and were more common at night than during the day. Growls also occurred more often during the spawning period than the pre-spawning period, while the trend for snaps was reversed. In a laboratory flume, sounds occurred when male lake trout were displaying spawning behaviors; growls when males were quivering and parallel swimming, and snaps when males moved their jaw. Combining our results with the observation of possible sound production by spawning splake (Salvelinus fontinalis × Salvelinus namaycush hybrid), provides rare evidence for spawning-related sound production by a salmonid, or any other fish in the superorder Protacanthopterygii. Further characterization of these sounds could be useful for lake trout assessment, restoration, and control.

  6. Aeroacoustics. [analysis of properties of sound generated by aerodynamic forces

    NASA Technical Reports Server (NTRS)

    Goldstein, M., E.

    1974-01-01

    An analysis was conducted to determine the properties of sound generated by aerodynamic forces or motions originating in a flow, such as the unsteady aerodynamic forces on propellers or by turbulent flows around an aircraft. The acoustics of moving media are reviewed and mathematical models are developed. Lighthill's acoustic analogy and the application to turbulent flows are analyzed. The effects of solid boundaries are calculated. Theories based on the solution of linearized vorticity and acoustic field equations are explained. The effects of nonuniform mean flow on the generation of sound are reported.

  7. Acoustic signatures of sound source-tract coupling.

    PubMed

    Arneodo, Ezequiel M; Perl, Yonatan Sanz; Mindlin, Gabriel B

    2011-04-01

    Birdsong is a complex behavior, which results from the interaction between a nervous system and a biomechanical peripheral device. While much has been learned about how complex sounds are generated in the vocal organ, little has been learned about the signature on the vocalizations of the nonlinear effects introduced by the acoustic interactions between a sound source and the vocal tract. The variety of morphologies among bird species makes birdsong a most suitable model to study phenomena associated to the production of complex vocalizations. Inspired by the sound production mechanisms of songbirds, in this work we study a mathematical model of a vocal organ, in which a simple sound source interacts with a tract, leading to a delay differential equation. We explore the system numerically, and by taking it to the weakly nonlinear limit, we are able to examine its periodic solutions analytically. By these means we are able to explore the dynamics of oscillatory solutions of a sound source-tract coupled system, which are qualitatively different from those of a sound source-filter model of a vocal organ. Nonlinear features of the solutions are proposed as the underlying mechanisms of observed phenomena in birdsong, such as unilaterally produced "frequency jumps," enhancement of resonances, and the shift of the fundamental frequency observed in heliox experiments. ©2011 American Physical Society

  8. Acoustic signatures of sound source-tract coupling

    PubMed Central

    Arneodo, Ezequiel M.; Perl, Yonatan Sanz; Mindlin, Gabriel B.

    2014-01-01

    Birdsong is a complex behavior, which results from the interaction between a nervous system and a biomechanical peripheral device. While much has been learned about how complex sounds are generated in the vocal organ, little has been learned about the signature on the vocalizations of the nonlinear effects introduced by the acoustic interactions between a sound source and the vocal tract. The variety of morphologies among bird species makes birdsong a most suitable model to study phenomena associated to the production of complex vocalizations. Inspired by the sound production mechanisms of songbirds, in this work we study a mathematical model of a vocal organ, in which a simple sound source interacts with a tract, leading to a delay differential equation. We explore the system numerically, and by taking it to the weakly nonlinear limit, we are able to examine its periodic solutions analytically. By these means we are able to explore the dynamics of oscillatory solutions of a sound source-tract coupled system, which are qualitatively different from those of a sound source-filter model of a vocal organ. Nonlinear features of the solutions are proposed as the underlying mechanisms of observed phenomena in birdsong, such as unilaterally produced “frequency jumps,” enhancement of resonances, and the shift of the fundamental frequency observed in heliox experiments. PMID:21599213

  9. Separation of non-stationary multi-source sound field based on the interpolated time-domain equivalent source method

    NASA Astrophysics Data System (ADS)

    Bi, Chuan-Xing; Geng, Lin; Zhang, Xiao-Zheng

    2016-05-01

    In the sound field with multiple non-stationary sources, the measured pressure is the sum of the pressures generated by all sources, and thus cannot be used directly for studying the vibration and sound radiation characteristics of every source alone. This paper proposes a separation model based on the interpolated time-domain equivalent source method (ITDESM) to separate the pressure field belonging to every source from the non-stationary multi-source sound field. In the proposed method, ITDESM is first extended to establish the relationship between the mixed time-dependent pressure and all the equivalent sources distributed on every source with known location and geometry information, and all the equivalent source strengths at each time step are solved by an iterative solving process; then, the corresponding equivalent source strengths of one interested source are used to calculate the pressure field generated by that source alone. Numerical simulation of two baffled circular pistons demonstrates that the proposed method can be effective in separating the non-stationary pressure generated by every source alone in both time and space domains. An experiment with two speakers in a semi-anechoic chamber further evidences the effectiveness of the proposed method.

  10. Intensity-invariant coding in the auditory system.

    PubMed

    Barbour, Dennis L

    2011-11-01

    The auditory system faithfully represents sufficient details from sound sources such that downstream cognitive processes are capable of acting upon this information effectively even in the face of signal uncertainty, degradation or interference. This robust sound source representation leads to an invariance in perception vital for animals to interact effectively with their environment. Due to unique nonlinearities in the cochlea, sound representations early in the auditory system exhibit a large amount of variability as a function of stimulus intensity. In other words, changes in stimulus intensity, such as for sound sources at differing distances, create a unique challenge for the auditory system to encode sounds invariantly across the intensity dimension. This challenge and some strategies available to sensory systems to eliminate intensity as an encoding variable are discussed, with a special emphasis upon sound encoding. Copyright © 2011 Elsevier Ltd. All rights reserved.

  11. Introduction to Exploring Our Bodies

    ERIC Educational Resources Information Center

    Early Childhood Today, 2006

    2006-01-01

    The body is probably one of a child's first science experiments. Young children are curious about their own bodies and how they work. They love to explore how they move (and do not move), the sounds they makes, how they look, how different textures feel on their skin, even how it tastes when they suck their thumb. Activities suggested in this…

  12. Numerical Models for Sound Propagation in Long Spaces

    NASA Astrophysics Data System (ADS)

    Lai, Chenly Yuen Cheung

    Both reverberation time and steady-state sound field are the key elements for assessing the acoustic condition in an enclosed space. They affect the noise propagation, speech intelligibility, clarity index, and definition. Since the sound field in a long space is non diffuse, classical room acoustics theory does not apply in this situation. The ray tracing technique and the image source methods are two common models to fathom both reverberation time and steady-state sound field in long enclosures nowadays. Although both models can give an accurate estimate of reverberation times and steady-state sound field directly or indirectly, they often involve time-consuming calculations. In order to simplify the acoustic consideration, a theoretical formulation has been developed for predicting both steady-state sound fields and reverberation times in street canyons. The prediction model is further developed to predict the steady-state sound field in a long enclosure. Apart from the straight long enclosure, there are other variations such as a cross junction, a long enclosure with a T-intersection, an U-turn long enclosure. In the present study, an theoretical and experimental investigations were conducted to develop formulae for predicting reverberation times and steady-state sound fields in a junction of a street canyon and in a long enclosure with T-intersection. The theoretical models are validated by comparing the numerical predictions with published experimental results. The theoretical results are also compared with precise indoor measurements and large-scale outdoor experimental results. In all of previous acoustical studies related to long enclosure, most of the studies are focused on the monopole sound source. Besides non-directional noise source, many noise sources in long enclosure are dipole like, such as train noise and fan noise. In order to study the characteristics of directional noise sources, a review of available dipole source was conducted. A dipole was constructed which was subsequent used for experimental studies. In additional, a theoretical model was developed for predicting dipole sound fields. The theoretical model can be used to study the effect of a dipole source on the speech intelligibility in long enclosures.

  13. Sound source tracking device for telematic spatial sound field reproduction

    NASA Astrophysics Data System (ADS)

    Cardenas, Bruno

    This research describes an algorithm that localizes sound sources for use in telematic applications. The localization algorithm is based on amplitude differences between various channels of a microphone array of directional shotgun microphones. The amplitude differences will be used to locate multiple performers and reproduce their voices, which were recorded at close distance with lavalier microphones, spatially corrected using a loudspeaker rendering system. In order to track multiple sound sources in parallel the information gained from the lavalier microphones will be utilized to estimate the signal-to-noise ratio between each performer and the concurrent performers.

  14. A Corticothalamic Circuit Model for Sound Identification in Complex Scenes

    PubMed Central

    Otazu, Gonzalo H.; Leibold, Christian

    2011-01-01

    The identification of the sound sources present in the environment is essential for the survival of many animals. However, these sounds are not presented in isolation, as natural scenes consist of a superposition of sounds originating from multiple sources. The identification of a source under these circumstances is a complex computational problem that is readily solved by most animals. We present a model of the thalamocortical circuit that performs level-invariant recognition of auditory objects in complex auditory scenes. The circuit identifies the objects present from a large dictionary of possible elements and operates reliably for real sound signals with multiple concurrently active sources. The key model assumption is that the activities of some cortical neurons encode the difference between the observed signal and an internal estimate. Reanalysis of awake auditory cortex recordings revealed neurons with patterns of activity corresponding to such an error signal. PMID:21931668

  15. Device for recording the 20 Hz - 200 KHz sound frequency spectrum using teletransmission

    NASA Technical Reports Server (NTRS)

    Baciu, I.

    1974-01-01

    The device described consists of two distinct parts: (1) The sound pickup system consisting of the wide-frequency band condenser microphone which contains in the same assembly the frequency-modulated oscillator and the output stage. Being transistorized and small, this system can be easily moved, so that sounds can be picked up even in places that are difficult to reach with larger devices. (2) The receiving and recording part is separate and can be at a great distance from the sound pickup system. This part contains a 72 MHz input stage, a frequency changer that gives an intermediate frequency of 30 MHz and a multichannel analyzer coupled to an oscilloscope and a recorder.

  16. Speed of Sound in Metal Pipes: An Inexpensive Lab

    ERIC Educational Resources Information Center

    Huggins, Elisha

    2008-01-01

    Our favorite demonstration for sound waves is to set up a compressional pulse on a horizontally stretched Slinky[TM]. One can easily watch the pulse move back and forth at a speed of the order of one meter per second. Watching this demonstration, it occurred to us that the same thing might happen in a steel pipe if you hit the end of the pipe with…

  17. Auditory performance in an open sound field

    NASA Astrophysics Data System (ADS)

    Fluitt, Kim F.; Letowski, Tomasz; Mermagen, Timothy

    2003-04-01

    Detection and recognition of acoustic sources in an open field are important elements of situational awareness on the battlefield. They are affected by many technical and environmental conditions such as type of sound, distance to a sound source, terrain configuration, meteorological conditions, hearing capabilities of the listener, level of background noise, and the listener's familiarity with the sound source. A limited body of knowledge about auditory perception of sources located over long distances makes it difficult to develop models predicting auditory behavior on the battlefield. The purpose of the present study was to determine the listener's abilities to detect, recognize, localize, and estimate distances to sound sources from 25 to 800 m from the listing position. Data were also collected for meteorological conditions (wind direction and strength, temperature, atmospheric pressure, humidity) and background noise level for each experimental trial. Forty subjects (men and women, ages 18 to 25) participated in the study. Nine types of sounds were presented from six loudspeakers in random order; each series was presented four times. Partial results indicate that both detection and recognition declined at distances greater than approximately 200 m and distance estimation was grossly underestimated by listeners. Specific results will be presented.

  18. Evolutionary trends in directional hearing

    PubMed Central

    Carr, Catherine E.; Christensen-Dalsgaard, Jakob

    2016-01-01

    Tympanic hearing is a true evolutionary novelty that arose in parallel within early tetrapods. We propose that in these tetrapods, selection for sound localization in air acted upon pre-existing directionally sensitive brainstem circuits, similar to those in fishes. Auditory circuits in birds and lizards resemble this ancestral, directionally sensitive framework. Despite this anatomically similarity, coding of sound source location differs between birds and lizards. In birds, brainstem circuits compute sound location from interaural cues. Lizards, however, have coupled ears, and do not need to compute source location in the brain. Thus their neural processing of sound direction differs, although all show mechanisms for enhancing sound source directionality. Comparisons with mammals reveal similarly complex interactions between coding strategies and evolutionary history. PMID:27448850

  19. Acoustic-gravity waves in atmospheric and oceanic waveguides.

    PubMed

    Godin, Oleg A

    2012-08-01

    A theory of guided propagation of sound in layered, moving fluids is extended to include acoustic-gravity waves (AGWs) in waveguides with piecewise continuous parameters. The orthogonality of AGW normal modes is established in moving and motionless media. A perturbation theory is developed to quantify the relative significance of the gravity and fluid compressibility as well as sensitivity of the normal modes to variations in sound speed, flow velocity, and density profiles and in boundary conditions. Phase and group speeds of the normal modes are found to have certain universal properties which are valid for waveguides with arbitrary stratification. The Lamb wave is shown to be the only AGW normal mode that can propagate without dispersion in a layered medium.

  20. Enhancing Auditory Selective Attention Using a Visually Guided Hearing Aid.

    PubMed

    Kidd, Gerald

    2017-10-17

    Listeners with hearing loss, as well as many listeners with clinically normal hearing, often experience great difficulty segregating talkers in a multiple-talker sound field and selectively attending to the desired "target" talker while ignoring the speech from unwanted "masker" talkers and other sources of sound. This listening situation forms the classic "cocktail party problem" described by Cherry (1953) that has received a great deal of study over the past few decades. In this article, a new approach to improving sound source segregation and enhancing auditory selective attention is described. The conceptual design, current implementation, and results obtained to date are reviewed and discussed in this article. This approach, embodied in a prototype "visually guided hearing aid" (VGHA) currently used for research, employs acoustic beamforming steered by eye gaze as a means for improving the ability of listeners to segregate and attend to one sound source in the presence of competing sound sources. The results from several studies demonstrate that listeners with normal hearing are able to use an attention-based "spatial filter" operating primarily on binaural cues to selectively attend to one source among competing spatially distributed sources. Furthermore, listeners with sensorineural hearing loss generally are less able to use this spatial filter as effectively as are listeners with normal hearing especially in conditions high in "informational masking." The VGHA enhances auditory spatial attention for speech-on-speech masking and improves signal-to-noise ratio for conditions high in "energetic masking." Visual steering of the beamformer supports the coordinated actions of vision and audition in selective attention and facilitates following sound source transitions in complex listening situations. Both listeners with normal hearing and with sensorineural hearing loss may benefit from the acoustic beamforming implemented by the VGHA, especially for nearby sources in less reverberant sound fields. Moreover, guiding the beam using eye gaze can be an effective means of sound source enhancement for listening conditions where the target source changes frequently over time as often occurs during turn-taking in a conversation. http://cred.pubs.asha.org/article.aspx?articleid=2601621.

  1. Enhancing Auditory Selective Attention Using a Visually Guided Hearing Aid

    PubMed Central

    2017-01-01

    Purpose Listeners with hearing loss, as well as many listeners with clinically normal hearing, often experience great difficulty segregating talkers in a multiple-talker sound field and selectively attending to the desired “target” talker while ignoring the speech from unwanted “masker” talkers and other sources of sound. This listening situation forms the classic “cocktail party problem” described by Cherry (1953) that has received a great deal of study over the past few decades. In this article, a new approach to improving sound source segregation and enhancing auditory selective attention is described. The conceptual design, current implementation, and results obtained to date are reviewed and discussed in this article. Method This approach, embodied in a prototype “visually guided hearing aid” (VGHA) currently used for research, employs acoustic beamforming steered by eye gaze as a means for improving the ability of listeners to segregate and attend to one sound source in the presence of competing sound sources. Results The results from several studies demonstrate that listeners with normal hearing are able to use an attention-based “spatial filter” operating primarily on binaural cues to selectively attend to one source among competing spatially distributed sources. Furthermore, listeners with sensorineural hearing loss generally are less able to use this spatial filter as effectively as are listeners with normal hearing especially in conditions high in “informational masking.” The VGHA enhances auditory spatial attention for speech-on-speech masking and improves signal-to-noise ratio for conditions high in “energetic masking.” Visual steering of the beamformer supports the coordinated actions of vision and audition in selective attention and facilitates following sound source transitions in complex listening situations. Conclusions Both listeners with normal hearing and with sensorineural hearing loss may benefit from the acoustic beamforming implemented by the VGHA, especially for nearby sources in less reverberant sound fields. Moreover, guiding the beam using eye gaze can be an effective means of sound source enhancement for listening conditions where the target source changes frequently over time as often occurs during turn-taking in a conversation. Presentation Video http://cred.pubs.asha.org/article.aspx?articleid=2601621 PMID:29049603

  2. Relation of sound intensity and accuracy of localization.

    PubMed

    Farrimond, T

    1989-08-01

    Tests were carried out on 17 subjects to determine the accuracy of monaural sound localization when the head is not free to turn toward the sound source. Maximum accuracy of localization for a constant-volume sound source coincided with the position for maximum perceived intensity of the sound in the front quadrant. There was a tendency for sounds to be perceived more often as coming from a position directly toward the ear. That is, for sounds in the front quadrant, errors of localization tended to be predominantly clockwise (i.e., biased toward a line directly facing the ear). Errors for sounds occurring in the rear quadrant tended to be anticlockwise. The pinna's differential effect on sound intensity between front and rear quadrants would assist in identifying the direction of movement of objects, for example an insect, passing the ear.

  3. Source sparsity control of sound field reproduction using the elastic-net and the lasso minimizers.

    PubMed

    Gauthier, P-A; Lecomte, P; Berry, A

    2017-04-01

    Sound field reproduction is aimed at the reconstruction of a sound pressure field in an extended area using dense loudspeaker arrays. In some circumstances, sound field reproduction is targeted at the reproduction of a sound field captured using microphone arrays. Although methods and algorithms already exist to convert microphone array recordings to loudspeaker array signals, one remaining research question is how to control the spatial sparsity in the resulting loudspeaker array signals and what would be the resulting practical advantages. Sparsity is an interesting feature for spatial audio since it can drastically reduce the number of concurrently active reproduction sources and, therefore, increase the spatial contrast of the solution at the expense of a difference between the target and reproduced sound fields. In this paper, the application of the elastic-net cost function to sound field reproduction is compared to the lasso cost function. It is shown that the elastic-net can induce solution sparsity and overcomes limitations of the lasso: The elastic-net solves the non-uniqueness of the lasso solution, induces source clustering in the sparse solution, and provides a smoother solution within the activated source clusters.

  4. Structural parameter effect of porous material on sound absorption performance of double-resonance material

    NASA Astrophysics Data System (ADS)

    Fan, C.; Tian, Y.; Wang, Z. Q.; Nie, J. K.; Wang, G. K.; Liu, X. S.

    2017-06-01

    In view of the noise feature and service environment of urban power substations, this paper explores the idea of compound impedance, fills some porous sound-absorption material in the first resonance cavity of the double-resonance sound-absorption material, and designs a new-type of composite acoustic board. We conduct some acoustic characterizations according to the standard test of impedance tube, and research on the influence of assembly order, the thickness and area density of the filling material, and back cavity on material sound-absorption performance. The results show that the new-type of acoustic board consisting of aluminum fibrous material as inner structure, micro-porous board as outer structure, and polyester-filled space between them, has good sound-absorption performance for low frequency and full frequency noise. When the thickness, area density of filling material and thickness of back cavity increase, the sound absorption coefficient curve peak will move toward low frequency.

  5. Sound preference test in animal models of addicts and phobias.

    PubMed

    Soga, Ryo; Shiramatsu, Tomoyo I; Kanzaki, Ryohei; Takahashi, Hirokazu

    2016-08-01

    Biased or too strong preference for a particular object is often problematic, resulting in addiction and phobia. In animal models, alternative forced-choice tasks have been routinely used, but such preference test is far from daily situations that addicts or phobic are facing. In the present study, we developed a behavioral assay to evaluate the preference of sounds in rodents. In the assay, several sounds were presented according to the position of free-moving rats, and quantified the sound preference based on the behavior. A particular tone was paired with microstimulation to the ventral tegmental area (VTA), which plays central roles in reward processing, to increase sound preference. The behaviors of rats were logged during the classical conditioning for six days. Consequently, some behavioral indices suggest that rats search for the conditioned sound. Thus, our data demonstrated that quantitative evaluation of preference in the behavioral assay is feasible.

  6. Transfer of knowledge from sound quality measurement to noise impact evaluation

    NASA Astrophysics Data System (ADS)

    Genuit, Klaus

    2004-05-01

    It is well known that the measurement and analysis of sound quality requires a complex procedure with consideration of the physical, psychoacoustical and psychological aspects of sound. Sound quality cannot be described only by a simple value based on A-weighted sound pressure level measurements. The A-weighted sound pressure level is sufficient to predict the probabilty that the human ear could be damaged by sound but the A-weighted level is not the correct descriptor for the annoyance of a complex sound situation given by several different sound events at different and especially moving positions (soundscape). On the one side, the consideration of the spectral distribution and the temporal pattern (psychoacoustics) is requested and, on the other side, the subjective attitude with respect to the sound situation, the expectation and experience of the people (psychology) have to be included in context with the complete noise impact evaluation. This paper describes applications of the newest methods of sound quality measurements-as it is well introduced at the car manufacturers-based on artifical head recordings and signal processing comparable to the human hearing used in noisy environments like community/traffic noise.

  7. Sound quality indicators for urban places in Paris cross-validated by Milan data.

    PubMed

    Ricciardi, Paola; Delaitre, Pauline; Lavandier, Catherine; Torchia, Francesca; Aumond, Pierre

    2015-10-01

    A specific smartphone application was developed to collect perceptive and acoustic data in Paris. About 3400 questionnaires were analyzed, regarding the global sound environment characterization, the perceived loudness of some emergent sources and the presence time ratio of sources that do not emerge from the background. Sound pressure level was recorded each second from the mobile phone's microphone during a 10-min period. The aim of this study is to propose indicators of urban sound quality based on linear regressions with perceptive variables. A cross validation of the quality models extracted from Paris data was carried out by conducting the same survey in Milan. The proposed sound quality general model is correlated with the real perceived sound quality (72%). Another model without visual amenity and familiarity is 58% correlated with perceived sound quality. In order to improve the sound quality indicator, a site classification was performed by Kohonen's Artificial Neural Network algorithm, and seven specific class models were developed. These specific models attribute more importance on source events and are slightly closer to the individual data than the global model. In general, the Parisian models underestimate the sound quality of Milan environments assessed by Italian people.

  8. Virtual environment display for a 3D audio room simulation

    NASA Technical Reports Server (NTRS)

    Chapin, William L.; Foster, Scott H.

    1992-01-01

    The development of a virtual environment simulation system integrating a 3D acoustic audio model with an immersive 3D visual scene is discussed. The system complements the acoustic model and is specified to: allow the listener to freely move about the space, a room of manipulable size, shape, and audio character, while interactively relocating the sound sources; reinforce the listener's feeling of telepresence in the acoustical environment with visual and proprioceptive sensations; enhance the audio with the graphic and interactive components, rather than overwhelm or reduce it; and serve as a research testbed and technology transfer demonstration. The hardware/software design of two demonstration systems, one installed and one portable, are discussed through the development of four iterative configurations.

  9. LightLeaves: computer controlled kinetic reflection hologram installation and a brief discussion of earlier work

    NASA Astrophysics Data System (ADS)

    Connors Chen, Betsy

    2013-02-01

    LightLeaves is an installation combining leaf shaped, white light reflection holograms of landscape images with a special kinetic lighting device that houses a lamp and moving leaf shaped masks. The masks are controlled by an Arduino microcontroller and servomotors that position the masks in front of the illumination source of the holograms. The work is the most recent in a long series of landscapes that combine multi-hologram installations with computer controlled devices that play with the motion of the holograms, the light, sound or other elements in the work. LightLeaves was first exhibited at the Peabody Essex Museum in Salem, Massachusetts in a show titled "Eye Spy: Playing with Perception".

  10. Effects of Soundscapes on Perceived Crowding and Encounter Norms

    NASA Astrophysics Data System (ADS)

    Kim, Sang-Oh; Shelby, Bo

    2011-07-01

    Soundscapes in recreation settings are becoming an important issue, but there are few studies of the effects of sounds on recreation experiences, especially crowding perceptions and encounter norms. This study compared effects of six types of sounds (an airplane, a truck engine, children playing, birds, water, and a control) on perceived crowding (PC) and encounter norms for hikers. Data were collected from 47 college students through lab experiments using simulated images, with moving hikers inserted in the original photo taken in the Jungmeori area of Mudeungsan Provincial Park in Korea. Overall, the motor-made sounds of the airplane and truck engine increased PC and decreased acceptability ratings, and the natural sounds of birds and water decreased PC and increased acceptability ratings. Ratings of the sound of children playing were similar to those in the control (i.e., no sound). In addition, as numbers of hikers increased, the overall effects of sounds decreased, and there were few significant differences in PC or acceptability ratings at the highest encounter levels. Theoretical and methodological implications are discussed.

  11. Effects of soundscapes on perceived crowding and encounter norms.

    PubMed

    Kim, Sang-Oh; Shelby, Bo

    2011-07-01

    Soundscapes in recreation settings are becoming an important issue, but there are few studies of the effects of sounds on recreation experiences, especially crowding perceptions and encounter norms. This study compared effects of six types of sounds (an airplane, a truck engine, children playing, birds, water, and a control) on perceived crowding (PC) and encounter norms for hikers. Data were collected from 47 college students through lab experiments using simulated images, with moving hikers inserted in the original photo taken in the Jungmeori area of Mudeungsan Provincial Park in Korea. Overall, the motor-made sounds of the airplane and truck engine increased PC and decreased acceptability ratings, and the natural sounds of birds and water decreased PC and increased acceptability ratings. Ratings of the sound of children playing were similar to those in the control (i.e., no sound). In addition, as numbers of hikers increased, the overall effects of sounds decreased, and there were few significant differences in PC or acceptability ratings at the highest encounter levels. Theoretical and methodological implications are discussed.

  12. Development of an ICT-Based Air Column Resonance Learning Media

    NASA Astrophysics Data System (ADS)

    Purjiyanta, Eka; Handayani, Langlang; Marwoto, Putut

    2016-08-01

    Commonly, the sound source used in the air column resonance experiment is the tuning fork having disadvantage of unoptimal resonance results due to the sound produced which is getting weaker. In this study we made tones with varying frequency using the Audacity software which were, then, stored in a mobile phone as a source of sound. One advantage of this sound source is the stability of the resulting sound enabling it to produce the same powerful sound. The movement of water in a glass tube mounted on the tool resonance and the tone sound that comes out from the mobile phone were recorded by using a video camera. Sound resonances recorded were first, second, and third resonance, for each tone frequency mentioned. The resulting sound stays longer, so it can be used for the first, second, third and next resonance experiments. This study aimed to (1) explain how to create tones that can substitute tuning forks sound used in air column resonance experiments, (2) illustrate the sound wave that occurred in the first, second, and third resonance in the experiment, and (3) determine the speed of sound in the air. This study used an experimental method. It was concluded that; (1) substitute tones of a tuning fork sound can be made by using the Audacity software; (2) the form of sound waves that occured in the first, second, and third resonance in the air column resonance can be drawn based on the results of video recording of the air column resonance; and (3) based on the experiment result, the speed of sound in the air is 346.5 m/s, while based on the chart analysis with logger pro software, the speed of sound in the air is 343.9 ± 0.3171 m/s.

  13. Migration patterns of post-spawning Pacific herring in a subarctic sound

    NASA Astrophysics Data System (ADS)

    Bishop, Mary Anne; Eiler, John H.

    2018-01-01

    Understanding the distribution of Pacific herring (Clupea pallasii) can be challenging because spawning, feeding and overwintering may take place in different areas separated by 1000s of kilometers. Along the northern Gulf of Alaska, Pacific herring movements after spring spawning are largely unknown. During the fall and spring, herring have been seen moving from the Gulf of Alaska into Prince William Sound, a large embayment, suggesting that fish spawning in the Sound migrate out into the Gulf of Alaska. We acoustic-tagged 69 adult herring on spawning grounds in Prince William Sound during April 2013 to determine seasonal migratory patterns. We monitored departures from the spawning grounds as well as herring arrivals and movements between the major entrances connecting Prince William Sound and the Gulf of Alaska. Departures of herring from the spawning grounds coincided with cessation of major spawning events in the immediate area. After spawning, 43 of 69 tagged herring (62%) moved to the entrances of Prince William Sound over a span of 104 d, although most fish arrived within 10 d of their departure from the spawning grounds. A large proportion remained in these areas until mid-June, most likely foraging on the seasonal bloom of large, Neocalanus copepods. Pulses of tagged herring detected during September and October at Montague Strait suggest that some herring returned from the Gulf of Alaska. Intermittent detections at Montague Strait and the Port Bainbridge passages from September through early January (when the transmitters expired) indicate that herring schools are highly mobile and are overwintering in this area. The pattern of detections at the entrances to Prince William Sound suggest that some herring remain in the Gulf of Alaska until late winter. The results of this study confirm the connectivity between local herring stocks in Prince William Sound and the Gulf of Alaska.

  14. How the owl tracks its prey – II

    PubMed Central

    Takahashi, Terry T.

    2010-01-01

    Barn owls can capture prey in pitch darkness or by diving into snow, while homing in on the sounds made by their prey. First, the neural mechanisms by which the barn owl localizes a single sound source in an otherwise quiet environment will be explained. The ideas developed for the single source case will then be expanded to environments in which there are multiple sound sources and echoes – environments that are challenging for humans with impaired hearing. Recent controversies regarding the mechanisms of sound localization will be discussed. Finally, the case in which both visual and auditory information are available to the owl will be considered. PMID:20889819

  15. Design of laser monitoring and sound localization system

    NASA Astrophysics Data System (ADS)

    Liu, Yu-long; Xu, Xi-ping; Dai, Yu-ming; Qiao, Yang

    2013-08-01

    In this paper, a novel design of laser monitoring and sound localization system is proposed. It utilizes laser to monitor and locate the position of the indoor conversation. In China most of the laser monitors no matter used in labor in an instrument uses photodiode or phototransistor as a detector at present. At the laser receivers of those facilities, light beams are adjusted to ensure that only part of the window in photodiodes or phototransistors received the beams. The reflection would deviate from its original path because of the vibration of the detected window, which would cause the changing of imaging spots in photodiode or phototransistor. However, such method is limited not only because it could bring in much stray light in receivers but also merely single output of photocurrent could be obtained. Therefore a new method based on quadrant detector is proposed. It utilizes the relation of the optical integral among quadrants to locate the position of imaging spots. This method could eliminate background disturbance and acquired two-dimensional spots vibrating data pacifically. The principle of this whole system could be described as follows. Collimated laser beams are reflected from vibrate-window caused by the vibration of sound source. Therefore reflected beams are modulated by vibration source. Such optical signals are collected by quadrant detectors and then are processed by photoelectric converters and corresponding circuits. Speech signals are eventually reconstructed. In addition, sound source localization is implemented by the means of detecting three different reflected light sources simultaneously. Indoor mathematical models based on the principle of Time Difference Of Arrival (TDOA) are established to calculate the twodimensional coordinate of sound source. Experiments showed that this system is able to monitor the indoor sound source beyond 15 meters with a high quality of speech reconstruction and to locate the sound source position accurately.

  16. Assessment of Hydroacoustic Propagation Using Autonomous Hydrophones in the Scotia Sea

    DTIC Science & Technology

    2010-09-01

    Award No. DE-AI52-08NA28654 Proposal No. BAA08-36 ABSTRACT The remote area of the Atlantic Ocean near the Antarctic Peninsula and the South...hydroacoustic blind spot. To investigate the sound propagation and interferences affected by these landmasses in the vicinity of the Antarctic polar...from large icebergs (near-surface sources) were utilized as natural sound sources. Surface sound sources, e.g., ice-related events, tend to suffer less

  17. Active control of noise on the source side of a partition to increase its sound isolation

    NASA Astrophysics Data System (ADS)

    Tarabini, Marco; Roure, Alain; Pinhede, Cedric

    2009-03-01

    This paper describes a local active noise control system that virtually increases the sound isolation of a dividing wall by means of a secondary source array. With the proposed method, sound pressure on the source side of the partition is reduced using an array of loudspeakers that generates destructive interference on the wall surface, where an array of error microphones is placed. The reduction of sound pressure on the incident side of the wall is expected to decrease the sound radiated into the contiguous room. The method efficiency was experimentally verified by checking the insertion loss of the active noise control system; in order to investigate the possibility of using a large number of actuators, a decentralized FXLMS control algorithm was used. Active control performances and stability were tested with different array configurations, loudspeaker directivities and enclosure characteristics (sound source position and absorption coefficient). The influence of all these parameters was investigated with the factorial design of experiments. The main outcome of the experimental campaign was that the insertion loss produced by the secondary source array, in the 50-300 Hz frequency range, was close to 10 dB. In addition, the analysis of variance showed that the active noise control performance can be optimized with a proper choice of the directional characteristics of the secondary source and the distance between loudspeakers and error microphones.

  18. The low-frequency sound power measuring technique for an underwater source in a non-anechoic tank

    NASA Astrophysics Data System (ADS)

    Zhang, Yi-Ming; Tang, Rui; Li, Qi; Shang, Da-Jing

    2018-03-01

    In order to determine the radiated sound power of an underwater source below the Schroeder cut-off frequency in a non-anechoic tank, a low-frequency extension measuring technique is proposed. This technique is based on a unique relationship between the transmission characteristics of the enclosed field and those of the free field, which can be obtained as a correction term based on previous measurements of a known simple source. The radiated sound power of an unknown underwater source in the free field can thereby be obtained accurately from measurements in a non-anechoic tank. To verify the validity of the proposed technique, a mathematical model of the enclosed field is established using normal-mode theory, and the relationship between the transmission characteristics of the enclosed and free fields is obtained. The radiated sound power of an underwater transducer source is tested in a glass tank using the proposed low-frequency extension measuring technique. Compared with the free field, the radiated sound power level of the narrowband spectrum deviation is found to be less than 3 dB, and the 1/3 octave spectrum deviation is found to be less than 1 dB. The proposed testing technique can be used not only to extend the low-frequency applications of non-anechoic tanks, but also for measurement of radiated sound power from complicated sources in non-anechoic tanks.

  19. Modeling and analysis of secondary sources coupling for active sound field reduction in confined spaces

    NASA Astrophysics Data System (ADS)

    Montazeri, Allahyar; Taylor, C. James

    2017-10-01

    This article addresses the coupling of acoustic secondary sources in a confined space in a sound field reduction framework. By considering the coupling of sources in a rectangular enclosure, the set of coupled equations governing its acoustical behavior are solved. The model obtained in this way is used to analyze the behavior of multi-input multi-output (MIMO) active sound field control (ASC) systems, where the coupling of sources cannot be neglected. In particular, the article develops the analytical results to analyze the effect of coupling of an array of secondary sources on the sound pressure levels inside an enclosure, when an array of microphones is used to capture the acoustic characteristics of the enclosure. The results are supported by extensive numerical simulations showing how coupling of loudspeakers through acoustic modes of the enclosure will change the strength and hence the driving voltage signal applied to the secondary loudspeakers. The practical significance of this model is to provide a better insight on the performance of the sound reproduction/reduction systems in confined spaces when an array of loudspeakers and microphones are placed in a fraction of wavelength of the excitation signal to reduce/reproduce the sound field. This is of particular importance because the interaction of different sources affects their radiation impedance depending on the electromechanical properties of the loudspeakers.

  20. Consistent modelling of wind turbine noise propagation from source to receiver.

    PubMed

    Barlas, Emre; Zhu, Wei Jun; Shen, Wen Zhong; Dag, Kaya O; Moriarty, Patrick

    2017-11-01

    The unsteady nature of wind turbine noise is a major reason for annoyance. The variation of far-field sound pressure levels is not only caused by the continuous change in wind turbine noise source levels but also by the unsteady flow field and the ground characteristics between the turbine and receiver. To take these phenomena into account, a consistent numerical technique that models the sound propagation from the source to receiver is developed. Large eddy simulation with an actuator line technique is employed for the flow modelling and the corresponding flow fields are used to simulate sound generation and propagation. The local blade relative velocity, angle of attack, and turbulence characteristics are input to the sound generation model. Time-dependent blade locations and the velocity between the noise source and receiver are considered within a quasi-3D propagation model. Long-range noise propagation of a 5 MW wind turbine is investigated. Sound pressure level time series evaluated at the source time are studied for varying wind speeds, surface roughness, and ground impedances within a 2000 m radius from the turbine.

  1. Consistent modelling of wind turbine noise propagation from source to receiver

    DOE PAGES

    Barlas, Emre; Zhu, Wei Jun; Shen, Wen Zhong; ...

    2017-11-28

    The unsteady nature of wind turbine noise is a major reason for annoyance. The variation of far-field sound pressure levels is not only caused by the continuous change in wind turbine noise source levels but also by the unsteady flow field and the ground characteristics between the turbine and receiver. To take these phenomena into account, a consistent numerical technique that models the sound propagation from the source to receiver is developed. Large eddy simulation with an actuator line technique is employed for the flow modelling and the corresponding flow fields are used to simulate sound generation and propagation. Themore » local blade relative velocity, angle of attack, and turbulence characteristics are input to the sound generation model. Time-dependent blade locations and the velocity between the noise source and receiver are considered within a quasi-3D propagation model. Long-range noise propagation of a 5 MW wind turbine is investigated. Sound pressure level time series evaluated at the source time are studied for varying wind speeds, surface roughness, and ground impedances within a 2000 m radius from the turbine.« less

  2. Consistent modelling of wind turbine noise propagation from source to receiver

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Barlas, Emre; Zhu, Wei Jun; Shen, Wen Zhong

    The unsteady nature of wind turbine noise is a major reason for annoyance. The variation of far-field sound pressure levels is not only caused by the continuous change in wind turbine noise source levels but also by the unsteady flow field and the ground characteristics between the turbine and receiver. To take these phenomena into account, a consistent numerical technique that models the sound propagation from the source to receiver is developed. Large eddy simulation with an actuator line technique is employed for the flow modelling and the corresponding flow fields are used to simulate sound generation and propagation. Themore » local blade relative velocity, angle of attack, and turbulence characteristics are input to the sound generation model. Time-dependent blade locations and the velocity between the noise source and receiver are considered within a quasi-3D propagation model. Long-range noise propagation of a 5 MW wind turbine is investigated. Sound pressure level time series evaluated at the source time are studied for varying wind speeds, surface roughness, and ground impedances within a 2000 m radius from the turbine.« less

  3. Drive-by large-region acoustic noise-source mapping via sparse beamforming tomography.

    PubMed

    Tuna, Cagdas; Zhao, Shengkui; Nguyen, Thi Ngoc Tho; Jones, Douglas L

    2016-10-01

    Environmental noise is a risk factor for human physical and mental health, demanding an efficient large-scale noise-monitoring scheme. The current technology, however, involves extensive sound pressure level (SPL) measurements at a dense grid of locations, making it impractical on a city-wide scale. This paper presents an alternative approach using a microphone array mounted on a moving vehicle to generate two-dimensional acoustic tomographic maps that yield the locations and SPLs of the noise-sources sparsely distributed in the neighborhood traveled by the vehicle. The far-field frequency-domain delay-and-sum beamforming output power values computed at multiple locations as the vehicle drives by are used as tomographic measurements. The proposed method is tested with acoustic data collected by driving an electric vehicle with a rooftop-mounted microphone array along a straight road next to a large open field, on which various pre-recorded noise-sources were produced by a loudspeaker at different locations. The accuracy of the tomographic imaging results demonstrates the promise of this approach for rapid, low-cost environmental noise-monitoring.

  4. Broad band sound from wind turbine generators

    NASA Technical Reports Server (NTRS)

    Hubbard, H. H.; Shepherd, K. P.; Grosveld, F. W.

    1981-01-01

    Brief descriptions are given of the various types of large wind turbines and their sound characteristics. Candidate sources of broadband sound are identified and are rank ordered for a large upwind configuration wind turbine generator for which data are available. The rotor is noted to be the main source of broadband sound which arises from inflow turbulence and from the interactions of the turbulent boundary layer on the blade with its trailing edge. Sound is radiated about equally in all directions but the refraction effects of the wind produce an elongated contour pattern in the downwind direction.

  5. Effects of sound source directivity on auralizations

    NASA Astrophysics Data System (ADS)

    Sheets, Nathan W.; Wang, Lily M.

    2002-05-01

    Auralization, the process of rendering audible the sound field in a simulated space, is a useful tool in the design of acoustically sensitive spaces. The auralization depends on the calculation of an impulse response between a source and a receiver which have certain directional behavior. Many auralizations created to date have used omnidirectional sources; the effects of source directivity on auralizations is a relatively unexplored area. To examine if and how the directivity of a sound source affects the acoustical results obtained from a room, we used directivity data for three sources in a room acoustic modeling program called Odeon. The three sources are: violin, piano, and human voice. The results from using directional data are compared to those obtained using omnidirectional source behavior, both through objective measure calculations and subjective listening tests.

  6. Development of a directivity-controlled piezoelectric transducer for sound reproduction

    NASA Astrophysics Data System (ADS)

    Bédard, Magella; Berry, Alain

    2008-04-01

    Present sound reproduction systems do not attempt to simulate the spatial radiation of musical instruments, or sound sources in general, even though the spatial directivity has a strong impact on the psychoacoustic experience. A transducer consisting of 4 piezoelectric elemental sources made from curved PVDF films is used to generate a target directivity pattern in the horizontal plane, in the frequency range of 5-20 kHz. The vibratory and acoustical response of an elemental source is addressed, both theoretically and experimentally. Two approaches to synthesize the input signals to apply to each elemental source are developed in order to create a prescribed, frequency-dependent acoustic directivity. The circumferential Fourier decomposition of the target directivity provides a compromise between the magnitude and the phase reconstruction, whereas the minimization of a quadratic error criterion provides a best magnitude reconstruction. This transducer can improve sound reproduction by introducing the spatial radiation aspect of the original source at high frequency.

  7. Callback response of dugongs to conspecific chirp playbacks.

    PubMed

    Ichikawa, Kotaro; Akamatsu, Tomonari; Shinke, Tomio; Adulyanukosol, Kanjana; Arai, Nobuaki

    2011-06-01

    Dugongs (Dugong dugon) produce bird-like calls such as chirps and trills. The vocal responses of dugongs to playbacks of several acoustic stimuli were investigated. Animals were exposed to four different playback stimuli: a recorded chirp from a wild dugong, a synthesized down-sweep sound, a synthesized constant-frequency sound, and silence. Wild dugongs vocalized more frequently after playback of broadcast chirps than that after constant-frequency sounds or silence. The down-sweep sound also elicited more vocal responses than did silence. No significant difference was found between the broadcast chirps and the down-sweep sound. The ratio of wild dugong chirps to all calls and the dominant frequencies of the wild dugong calls were significantly higher during playbacks of broadcast chirps, down-sweep sounds, and constant-frequency sounds than during those of silence. The source level and duration of dugong chirps increased significantly as signaling distance increased. No significant correlation was found between signaling distance and the source level of trills. These results show that dugongs vocalize to playbacks of frequency-modulated signals and suggest that the source level of dugong chirps may be manipulated to compensate for transmission loss between the source and receiver. This study provides the first behavioral observations revealing the function of dugong chirps. © 2011 Acoustical Society of America

  8. Atmospheric Propagation

    NASA Technical Reports Server (NTRS)

    Embleton, Tony F. W.; Daigle, Gilles A.

    1991-01-01

    Reviewed here is the current state of knowledge with respect to each basic mechanism of sound propagation in the atmosphere and how each mechanism changes the spectral or temporal characteristics of the sound received at a distance from the source. Some of the basic processes affecting sound wave propagation which are present in any situation are discussed. They are geometrical spreading, molecular absorption, and turbulent scattering. In geometrical spreading, sound levels decrease with increasing distance from the source; there is no frequency dependence. In molecular absorption, sound energy is converted into heat as the sound wave propagates through the air; there is a strong dependence on frequency. In turbulent scattering, local variations in wind velocity and temperature induce fluctuations in phase and amplitude of the sound waves as they propagate through an inhomogeneous medium; there is a moderate dependence on frequency.

  9. How effectively do horizontal and vertical response strategies of long-finned pilot whales reduce sound exposure from naval sonar?

    PubMed

    Wensveen, Paul J; von Benda-Beckmann, Alexander M; Ainslie, Michael A; Lam, Frans-Peter A; Kvadsheim, Petter H; Tyack, Peter L; Miller, Patrick J O

    2015-05-01

    The behaviour of a marine mammal near a noise source can modulate the sound exposure it receives. We demonstrate that two long-finned pilot whales both surfaced in synchrony with consecutive arrivals of multiple sonar pulses. We then assess the effect of surfacing and other behavioural response strategies on the received cumulative sound exposure levels and maximum sound pressure levels (SPLs) by modelling realistic spatiotemporal interactions of a pilot whale with an approaching source. Under the propagation conditions of our model, some response strategies observed in the wild were effective in reducing received levels (e.g. movement perpendicular to the source's line of approach), but others were not (e.g. switching from deep to shallow diving; synchronous surfacing after maximum SPLs). Our study exemplifies how simulations of source-whale interactions guided by detailed observational data can improve our understanding about motivations behind behaviour responses observed in the wild (e.g., reducing sound exposure, prey movement). Copyright © 2015 Elsevier Ltd. All rights reserved.

  10. Difference in precedence effect between children and adults signifies development of sound localization abilities in complex listening tasks

    PubMed Central

    Litovsky, Ruth Y.; Godar, Shelly P.

    2010-01-01

    The precedence effect refers to the fact that humans are able to localize sound in reverberant environments, because the auditory system assigns greater weight to the direct sound (lead) than the later-arriving sound (lag). In this study, absolute sound localization was studied for single source stimuli and for dual source lead-lag stimuli in 4–5 year old children and adults. Lead-lag delays ranged from 5–100 ms. Testing was conducted in free field, with pink noise bursts emitted from loudspeakers positioned on a horizontal arc in the frontal field. Listeners indicated how many sounds were heard and the perceived location of the first- and second-heard sounds. Results suggest that at short delays (up to 10 ms), the lead dominates sound localization strongly at both ages, and localization errors are similar to those with single-source stimuli. At longer delays errors can be large, stemming from over-integration of the lead and lag, interchanging of perceived locations of the first-heard and second-heard sounds due to temporal order confusion, and dominance of the lead over the lag. The errors are greater for children than adults. Results are discussed in the context of maturation of auditory and non-auditory factors. PMID:20968369

  11. Simulated seal scarer sounds scare porpoises, but not seals: species-specific responses to 12 kHz deterrence sounds

    PubMed Central

    Hermannsen, Line; Beedholm, Kristian

    2017-01-01

    Acoustic harassment devices (AHD) or ‘seal scarers’ are used extensively, not only to deter seals from fisheries, but also as mitigation tools to deter marine mammals from potentially harmful sound sources, such as offshore pile driving. To test the effectiveness of AHDs, we conducted two studies with similar experimental set-ups on two key species: harbour porpoises and harbour seals. We exposed animals to 500 ms tone bursts at 12 kHz simulating that of an AHD (Lofitech), but with reduced output levels (source peak-to-peak level of 165 dB re 1 µPa). Animals were localized with a theodolite before, during and after sound exposures. In total, 12 sound exposures were conducted to porpoises and 13 exposures to seals. Porpoises were found to exhibit avoidance reactions out to ranges of 525 m from the sound source. Contrary to this, seal observations increased during sound exposure within 100 m of the loudspeaker. We thereby demonstrate that porpoises and seals respond very differently to AHD sounds. This has important implications for application of AHDs in multi-species habitats, as sound levels required to deter less sensitive species (seals) can lead to excessive and unwanted large deterrence ranges on more sensitive species (porpoises). PMID:28791155

  12. Characterizing large river sounds: Providing context for understanding the environmental effects of noise produced by hydrokinetic turbines

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bevelhimer, Mark S.; Deng, Z. Daniel; Scherelis, Constantin

    2016-01-01

    Underwater noise associated with the installation and operation of hydrokinetic turbines in rivers and tidal zones presents a potential environmental concern for fish and marine mammals. Comparing the spectral quality of sounds emitted by hydrokinetic turbines to natural and other anthropogenic sound sources is an initial step at understanding potential environmental impacts. Underwater recordings were obtained from passing vessels of different sizes and other underwater sound sources in both static and flowing waters. Static water measurements were taken in a lake with minimal background noise. Flowing water measurements were taken at a previously proposed deployment site for hydrokinetic turbines onmore » the Mississippi River, where the sound of flowing water is included in background measurements. The size of vessels measured ranged from a small fishing boat with a 60 HP outboard motor to an 18-unit barge train being pushed upstream by tugboat. As expected, large vessels with large engines created the highest sound levels, and when compared to the sound created by an operating HK turbine were many times greater. A comparison of sound levels from the same sources at different distances using both spherical and cylindrical sound attenuation functions suggests that spherical model results more closely approximate observed values.« less

  13. Feasibility of making sound power measurements in the NASA Langley V/STOL tunnel test section

    NASA Technical Reports Server (NTRS)

    Brooks, T. F.; Scheiman, J.; Silcox, R. J.

    1976-01-01

    Based on exploratory acoustic measurements in Langley's V/STOL wind tunnel, recommendations are made on the methodology for making sound power measurements of aircraft components in the closed tunnel test section. During airflow, tunnel self-noise and microphone flow-induced noise place restrictions on the amplitude and spectrum of the sound source to be measured. Models of aircraft components with high sound level sources, such as thrust engines and powered lift systems, seem likely candidates for acoustic testing.

  14. Varying sediment sources (Hudson Strait, Cumberland Sound, Baffin Bay) to the NW Labrador Sea slope between and during Heinrich events 0 to 4

    USGS Publications Warehouse

    Andrews, John T.; Barber, D.C.; Jennings, A.E.; Eberl, D.D.; Maclean, B.; Kirby, M.E.; Stoner, J.S.

    2012-01-01

    Core HU97048-007PC was recovered from the continental Labrador Sea slope at a water depth of 945 m, 250 km seaward from the mouth of Cumberland Sound, and 400 km north of Hudson Strait. Cumberland Sound is a structural trough partly floored by Cretaceous mudstones and Paleozoic carbonates. The record extends from ∼10 to 58 ka. On-board logging revealed a complex series of lithofacies, including buff-colored detrital carbonate-rich sediments [Heinrich (H)-events] frequently bracketed by black facies. We investigate the provenance of these facies using quantitative X-ray diffraction on drill-core samples from Paleozoic and Cretaceous bedrock from the SE Baffin Island Shelf, and on the < 2-mm sediment fraction in a transect of five cores from Cumberland Sound to the NW Labrador Sea. A sediment unmixing program was used to discriminate between sediment sources, which included dolomite-rich sediments from Baffin Bay, calcite-rich sediments from Hudson Strait and discrete sources from Cumberland Sound. Results indicated that the bulk of the sediment was derived from Cumberland Sound, but Baffin Bay contributed to sediments coeval with H-0 (Younger Dryas), whereas Hudson Strait was the source during H-events 1–4. Contributions from the Cretaceous outcrops within Cumberland Sound bracket H-events, thus both leading and lagging Hudson Strait-sourced H-events.

  15. Peripheral mechanisms for vocal production in birds - differences and similarities to human speech and singing.

    PubMed

    Riede, Tobias; Goller, Franz

    2010-10-01

    Song production in songbirds is a model system for studying learned vocal behavior. As in humans, bird phonation involves three main motor systems (respiration, vocal organ and vocal tract). The avian respiratory mechanism uses pressure regulation in air sacs to ventilate a rigid lung. In songbirds sound is generated with two independently controlled sound sources, which reside in a uniquely avian vocal organ, the syrinx. However, the physical sound generation mechanism in the syrinx shows strong analogies to that in the human larynx, such that both can be characterized as myoelastic-aerodynamic sound sources. Similarities include active adduction and abduction, oscillating tissue masses which modulate flow rate through the organ and a layered structure of the oscillating tissue masses giving rise to complex viscoelastic properties. Differences in the functional morphology of the sound producing system between birds and humans require specific motor control patterns. The songbird vocal apparatus is adapted for high speed, suggesting that temporal patterns and fast modulation of sound features are important in acoustic communication. Rapid respiratory patterns determine the coarse temporal structure of song and maintain gas exchange even during very long songs. The respiratory system also contributes to the fine control of airflow. Muscular control of the vocal organ regulates airflow and acoustic features. The upper vocal tract of birds filters the sounds generated in the syrinx, and filter properties are actively adjusted. Nonlinear source-filter interactions may also play a role. The unique morphology and biomechanical system for sound production in birds presents an interesting model for exploring parallels in control mechanisms that give rise to highly convergent physical patterns of sound generation. More comparative work should provide a rich source for our understanding of the evolution of complex sound producing systems. Copyright © 2009 Elsevier Inc. All rights reserved.

  16. The auditory P50 component to onset and offset of sound

    PubMed Central

    Pratt, Hillel; Starr, Arnold; Michalewski, Henry J.; Bleich, Naomi; Mittelman, Nomi

    2008-01-01

    Objective: The auditory Event-Related Potentials (ERP) component P50 to sound onset and offset have been reported to be similar, but their magnetic homologue has been reported absent to sound offset. We compared the spatio-temporal distribution of cortical activity during P50 to sound onset and offset, without confounds of spectral change. Methods: ERPs were recorded in response to onsets and offsets of silent intervals of 0.5 s (gaps) appearing randomly in otherwise continuous white noise and compared to ERPs to randomly distributed click pairs with half second separation presented in silence. Subjects were awake and distracted from the stimuli by reading a complicated text. Measures of P50 included peak latency and amplitude, as well as source current density estimates to the clicks and sound onsets and offsets. Results P50 occurred in response to noise onsets and to clicks, while to noise offset it was absent. Latency of P50 was similar to noise onset (56 msec) and to clicks (53 msec). Sources of P50 to noise onsets and clicks included bilateral superior parietal areas. In contrast, noise offsets activated left inferior temporal and occipital areas at the time of P50. Source current density was significantly higher to noise onset than offset in the vicinity of the temporo-parietal junction. Conclusions: P50 to sound offset is absent compared to the distinct P50 to sound onset and to clicks, at different intracranial sources. P50 to stimulus onset and to clicks appears to reflect preattentive arousal by a new sound in the scene. Sound offset does not involve a new sound and hence the absent P50. Significance: Stimulus onset activates distinct early cortical processes that are absent to offset. PMID:18055255

  17. Blind separation of incoherent and spatially disjoint sound sources

    NASA Astrophysics Data System (ADS)

    Dong, Bin; Antoni, Jérôme; Pereira, Antonio; Kellermann, Walter

    2016-11-01

    Blind separation of sound sources aims at reconstructing the individual sources which contribute to the overall radiation of an acoustical field. The challenge is to reach this goal using distant measurements when all sources are operating concurrently. The working assumption is usually that the sources of interest are incoherent - i.e. statistically orthogonal - so that their separation can be approached by decorrelating a set of simultaneous measurements, which amounts to diagonalizing the cross-spectral matrix. Principal Component Analysis (PCA) is traditionally used to this end. This paper reports two new findings in this context. First, a sufficient condition is established under which "virtual" sources returned by PCA coincide with true sources; it stipulates that the sources of interest should be not only incoherent but also spatially orthogonal. A particular case of this instance is met by spatially disjoint sources - i.e. with non-overlapping support sets. Second, based on this finding, a criterion that enforces both statistical and spatial orthogonality is proposed to blindly separate incoherent sound sources which radiate from disjoint domains. This criterion can be easily incorporated into acoustic imaging algorithms such as beamforming or acoustical holography to identify sound sources of different origins. The proposed methodology is validated on laboratory experiments. In particular, the separation of aeroacoustic sources is demonstrated in a wind tunnel.

  18. Virtual Manufacturing (la Fabrication virtuelle)

    DTIC Science & Technology

    1998-05-01

    with moving parts and subassemblies, • verification of product subcomponents and systems operations through kinematics studies, and • realism ...dimensions, parts moved in mechanism based directions, and realism of interaction is increased through use of sound, touch and other parameters. For the...direct converters from CAD systems. A simple cinematic package is also high on the requirement to be able to simulate motions as well as an interface to

  19. Investigation on flow oscillation modes and aero-acoustics generation mechanism in cavity

    NASA Astrophysics Data System (ADS)

    Yang, Dang-Guo; Lu, Bo; Cai, Jin-Sheng; Wu, Jun-Qiang; Qu, Kun; Liu, Jun

    2018-05-01

    Unsteady flow and multi-scale vortex transformation inside a cavity of L/D = 6 (ratio of length to depth) at Ma = 0.9 and 1.5 were studied using the numerical simulation method of modified delayed detached eddy simulation (DDES) in this paper. Aero-acoustic characteristics for the cavity at same flow conditions were obtained by the numerical method and 0.6 m by 0.6 m transonic and supersonic wind-tunnel experiments. The analysis on the computational and experimental results indicates that some vortex generates from flow separation in shear-layer over the cavity, and the vortex moves from forward to downward of the cavity at some velocity, and impingement of the vortex and the rear-wall of the cavity occurs. Some sound waves spread abroad to the cavity fore-wall, which induces some new vortex generation, and the vortex sheds, moves and impinges on the cavity rear-wall. New sound waves occur. The research results indicate that sound wave feedback created by the impingement of the shedding-vortices and rear cavity face leads to flow oscillations and noise generation inside the cavity. Analysis on aero-acoustic characteristics inside the cavity is feasible. The simulated self-sustained flow-oscillation modes and peak sound pressure on typical frequencies inside the cavity agree well with Rossiter’s and Heller’s predicated results. Moreover, the peak sound pressure occurs in the first and second flow-oscillation modes and most of sound energy focuses on the low-frequency region. Compared with subsonic speed (Ma = 0.9), aerodynamic noise is more intense at Ma = 1.5, which is induced by compression wave or shock wave in near region of fore and rear cavity face.

  20. Metals, organic compounds, and nutrients in Long Island Sound: sources, magnitudes, trends, and impacts

    USGS Publications Warehouse

    Mullaney, John R.; Varekamp, J.C.; MCElroy, A.E.; Brsslin, V.T.

    2014-01-01

    The main rivers that discharge into LIS are the East River in the west, the Housatonic and Connecticut rivers on the north, and the Thames River at the northeastern end of LIS, with the Quinnipiac and several other smaller rivers also coming in from Connecticut. The East River is a tidal strait that connects LIS with New York Harbor through the heart of the New York City metropolitan region. The Housatonic, Quinnipiac, Connecticut and Thames river basins drain agricultural, urban and industrial lands in a watershed that extends from Connecticut north to Canada. The Sound receives contaminants from many sources within and outside its contributing watershed, including direct discharges from coastal industries, wastewater treatment plants (WWTP), urban runoff, and atmospheric deposition. New England has a long history of industrial activity, with factories that once crowded its riverbanks and shores now having succumbed to economic forces that drove manufacturing overseas. Relict deposits with legacy pollutants in upland sediments persist and combine with modern runoff sources from an increasingly densely populated watershed, and continue to be a source of contaminants for LIS. While toxic exposure from legacy and active sources has diminished over the years as wastewater treatment has improved and industries closed or moved away, pockets of contamination still have consequences for many embayments and coves, particularly near urbanized areas of western LIS. Loading of nutrients and carbon have been of recent concern in LIS because of the extensive impacts observed since the mid-1980s. Excess nutrients not only create inhospitable conditions for higher forms of aquatic life through reduced oxygen levels and disrupting trophic dynamics, but also by altering the local biogeochemistry. As a result, the release of toxic substances into the water column may be enhanced in hypoxic waters, thus exerting a toxic effect or enhancing incorporation of toxic pollutants into the food we

  1. A New Mechanism of Sound Generation in Songbirds

    NASA Astrophysics Data System (ADS)

    Goller, Franz; Larsen, Ole N.

    1997-12-01

    Our current understanding of the sound-generating mechanism in the songbird vocal organ, the syrinx, is based on indirect evidence and theoretical treatments. The classical avian model of sound production postulates that the medial tympaniform membranes (MTM) are the principal sound generators. We tested the role of the MTM in sound generation and studied the songbird syrinx more directly by filming it endoscopically. After we surgically incapacitated the MTM as a vibratory source, zebra finches and cardinals were not only able to vocalize, but sang nearly normal song. This result shows clearly that the MTM are not the principal sound source. The endoscopic images of the intact songbird syrinx during spontaneous and brain stimulation-induced vocalizations illustrate the dynamics of syringeal reconfiguration before phonation and suggest a different model for sound production. Phonation is initiated by rostrad movement and stretching of the syrinx. At the same time, the syrinx is closed through movement of two soft tissue masses, the medial and lateral labia, into the bronchial lumen. Sound production always is accompanied by vibratory motions of both labia, indicating that these vibrations may be the sound source. However, because of the low temporal resolution of the imaging system, the frequency and phase of labial vibrations could not be assessed in relation to that of the generated sound. Nevertheless, in contrast to the previous model, these observations show that both labia contribute to aperture control and strongly suggest that they play an important role as principal sound generators.

  2. On the role of glottis-interior sources in the production of voiced sound.

    PubMed

    Howe, M S; McGowan, R S

    2012-02-01

    The voice source is dominated by aeroacoustic sources downstream of the glottis. In this paper an investigation is made of the contribution to voiced speech of secondary sources within the glottis. The acoustic waveform is ultimately determined by the volume velocity of air at the glottis, which is controlled by vocal fold vibration, pressure forcing from the lungs, and unsteady backreactions from the sound and from the supraglottal air jet. The theory of aerodynamic sound is applied to study the influence on the fine details of the acoustic waveform of "potential flow" added-mass-type glottal sources, glottis friction, and vorticity either in the glottis-wall boundary layer or in the portion of the free jet shear layer within the glottis. These sources govern predominantly the high frequency content of the sound when the glottis is near closure. A detailed analysis performed for a canonical, cylindrical glottis of rectangular cross section indicates that glottis-interior boundary/shear layer vortex sources and the surface frictional source are of comparable importance; the influence of the potential flow source is about an order of magnitude smaller. © 2012 Acoustical Society of America

  3. Calibration of International Space Station (ISS) Node 1 Vibro-Acoustic Model-Report 2

    NASA Technical Reports Server (NTRS)

    Zhang, Weiguo; Raveendra, Ravi

    2014-01-01

    Reported here is the capability of the Energy Finite Element Method (E-FEM) to predict the vibro-acoustic sound fields within the International Space Station (ISS) Node 1 and to compare the results with simulated leak sounds. A series of electronically generated structural ultrasonic noise sources were created in the pressure wall to emulate leak signals at different locations of the Node 1 STA module during its period of storage at Stennis Space Center (SSC). The exact sound source profiles created within the pressure wall at the source were unknown, but were estimated from the closest sensor measurement. The E-FEM method represents a reverberant sound field calculation, and of importance to this application is the requirement to correctly handle the direct field effect of the sound generation. It was also important to be able to compute the sound energy fields in the ultrasonic frequency range. This report demonstrates the capability of this technology as applied to this type of application.

  4. [Perception by teenagers and adults of the changed by amplitude sound sequences used in models of movement of the sound source].

    PubMed

    Andreeva, I G; Vartanian, I A

    2012-01-01

    The ability to evaluate direction of amplitude changes of sound stimuli was studied in adults and in the 11-12- and 15-16-year old teenagers. The stimuli representing sequences of fragments of the tone of 1 kHz, whose amplitude is changing with time, are used as model of approach and withdrawal of the sound sources. The 11-12-year old teenagers at estimation of direction of amplitude changes were shown to make the significantly higher number of errors as compared with two other examined groups, including those in repeated experiments. The structure of errors - the ratio of the portion of errors at estimation of increasing and decreasing by amplitude stimulus - turned out to be different in teenagers and in adults. The question is discussed about the effect of unspecific activation of the large hemisphere cortex in teenagers on processes if taking solution about the complex sound stimulus, including a possibility estimation of approach and withdrawal of the sound source.

  5. Interior and exterior sound field control using general two-dimensional first-order sources.

    PubMed

    Poletti, M A; Abhayapala, T D

    2011-01-01

    Reproduction of a given sound field interior to a circular loudspeaker array without producing an undesirable exterior sound field is an unsolved problem over a broadband of frequencies. At low frequencies, by implementing the Kirchhoff-Helmholtz integral using a circular discrete array of line-source loudspeakers, a sound field can be recreated within the array and produce no exterior sound field, provided that the loudspeakers have azimuthal polar responses with variable first-order responses which are a combination of a two-dimensional (2D) monopole and a radially oriented 2D dipole. This paper examines the performance of circular discrete arrays of line-source loudspeakers which also include a tangential dipole, providing general variable-directivity responses in azimuth. It is shown that at low frequencies, the tangential dipoles are not required, but that near and above the Nyquist frequency, the tangential dipoles can both improve the interior accuracy and reduce the exterior sound field. The additional dipoles extend the useful range of the array by around an octave.

  6. The silent base flow and the sound sources in a laminar jet.

    PubMed

    Sinayoko, Samuel; Agarwal, Anurag

    2012-03-01

    An algorithm to compute the silent base flow sources of sound in a jet is introduced. The algorithm is based on spatiotemporal filtering of the flow field and is applicable to multifrequency sources. It is applied to an axisymmetric laminar jet and the resulting sources are validated successfully. The sources are compared to those obtained from two classical acoustic analogies, based on quiescent and time-averaged base flows. The comparison demonstrates how the silent base flow sources shed light on the sound generation process. It is shown that the dominant source mechanism in the axisymmetric laminar jet is "shear-noise," which is a linear mechanism. The algorithm presented here could be applied to fully turbulent flows to understand the aerodynamic noise-generation mechanism. © 2012 Acoustical Society of America

  7. Sound Radiated by a Wave-Like Structure in a Compressible Jet

    NASA Technical Reports Server (NTRS)

    Golubev, V. V.; Prieto, A. F.; Mankbadi, R. R.; Dahl, M. D.; Hixon, R.

    2003-01-01

    This paper extends the analysis of acoustic radiation from the source model representing spatially-growing instability waves in a round jet at high speeds. Compared to previous work, a modified approach to the sound source modeling is examined that employs a set of solutions to linearized Euler equations. The sound radiation is then calculated using an integral surface method.

  8. Multilingual Vocabularies in Automatic Speech Recognition

    DTIC Science & Technology

    2000-08-01

    monolingual (a few thousands) is an obstacle to a full generalization of the inventories, then moved to the multilingual case. In the approach towards the...direction of language independence. In this monolingual experiment, we developed two types of unit sets for paper, we extend the method presented in [3...sound ji is not assimilated 3.2.1 Monolingual experiments to the corresponding sound in Spanish, but it is left apart as a The baseline model for English

  9. Measurement and Numerical Calculation of Force on a Particle in a Strong Acoustic Field Required for Levitation

    NASA Astrophysics Data System (ADS)

    Kozuka, Teruyuki; Yasui, Kyuichi; Tuziuti, Toru; Towata, Atsuya; Lee, Judy; Iida, Yasuo

    2009-07-01

    Using a standing-wave field generated between a sound source and a reflector, it is possible to trap small objects at nodes of the sound pressure distribution in air. In this study, a sound field generated under a flat or concave reflector was studied by both experimental measurement and numerical calculation. The calculated result agrees well with the experimental data. The maximum force generated between a sound source of 25.0 mm diameter and a concave reflector is 0.8 mN in the experiment. A steel ball of 2.0 mm in diameter was levitated in the sound field in air.

  10. Multivariate analysis of behavioural response experiments in humpback whales (Megaptera novaeangliae).

    PubMed

    Dunlop, Rebecca A; Noad, Michael J; Cato, Douglas H; Kniest, Eric; Miller, Patrick J O; Smith, Joshua N; Stokes, M Dale

    2013-03-01

    The behavioural response study (BRS) is an experimental design used by field biologists to determine the function and/or behavioural effects of conspecific, heterospecific or anthropogenic stimuli. When carrying out these studies in marine mammals it is difficult to make basic observations and achieve sufficient samples sizes because of the high cost and logistical difficulties. Rarely are other factors such as social context or the physical environment considered in the analysis because of these difficulties. This paper presents results of a BRS carried out in humpback whales to test the response of groups to one recording of conspecific social sounds and an artificially generated tone stimulus. Experiments were carried out in September/October 2004 and 2008 during the humpback whale southward migration along the east coast of Australia. In total, 13 'tone' experiments, 15 'social sound' experiments (using one recording of social sounds) and three silent controls were carried out over two field seasons. The results (using a mixed model statistical analysis) suggested that humpback whales responded differently to the two stimuli, measured by changes in course travelled and dive behaviour. Although the response to 'tones' was consistent, in that groups moved offshore and surfaced more often (suggesting an aversion to the stimulus), the response to 'social sounds' was highly variable and dependent upon the composition of the social group. The change in course and dive behaviour in response to 'tones' was found to be related to proximity to the source, the received signal level and signal-to-noise ratio (SNR). This study demonstrates that the behavioural responses of marine mammals to acoustic stimuli are complex. In order to tease out such multifaceted interactions, the number of replicates and factors measured must be sufficient for multivariate analysis.

  11. The effect of spatial auditory landmarks on ambulation.

    PubMed

    Karim, Adham M; Rumalla, Kavelin; King, Laurie A; Hullar, Timothy E

    2018-02-01

    The maintenance of balance and posture is a result of the collaborative efforts of vestibular, proprioceptive, and visual sensory inputs, but a fourth neural input, audition, may also improve balance. Here, we tested the hypothesis that auditory inputs function as environmental spatial landmarks whose effectiveness depends on sound localization ability during ambulation. Eight blindfolded normal young subjects performed the Fukuda-Unterberger test in three auditory conditions: silence, white noise played through headphones (head-referenced condition), and white noise played through a loudspeaker placed directly in front at 135 centimeters away from the ear at ear height (earth-referenced condition). For the earth-referenced condition, an additional experiment was performed where the effect of moving the speaker azimuthal position to 45, 90, 135, and 180° was tested. Subjects performed significantly better in the earth-referenced condition than in the head-referenced or silent conditions. Performance progressively decreased over the range from 0° to 135° but all subjects then improved slightly at the 180° compared to the 135° condition. These results suggest that presence of sound dramatically improves the ability to ambulate when vision is limited, but that sound sources must be located in the external environment in order to improve balance. This supports the hypothesis that they act by providing spatial landmarks against which head and body movement and orientation may be compared and corrected. Balance improvement in the azimuthal plane mirrors sensitivity to sound movement at similar positions, indicating that similar auditory mechanisms may underlie both processes. These results may help optimize the use of auditory cues to improve balance in particular patient populations. Copyright © 2017 Elsevier B.V. All rights reserved.

  12. Mouth and Voice: A Relationship between Visual and Auditory Preference in the Human Superior Temporal Sulcus

    PubMed Central

    2017-01-01

    Cortex in and around the human posterior superior temporal sulcus (pSTS) is known to be critical for speech perception. The pSTS responds to both the visual modality (especially biological motion) and the auditory modality (especially human voices). Using fMRI in single subjects with no spatial smoothing, we show that visual and auditory selectivity are linked. Regions of the pSTS were identified that preferred visually presented moving mouths (presented in isolation or as part of a whole face) or moving eyes. Mouth-preferring regions responded strongly to voices and showed a significant preference for vocal compared with nonvocal sounds. In contrast, eye-preferring regions did not respond to either vocal or nonvocal sounds. The converse was also true: regions of the pSTS that showed a significant response to speech or preferred vocal to nonvocal sounds responded more strongly to visually presented mouths than eyes. These findings can be explained by environmental statistics. In natural environments, humans see visual mouth movements at the same time as they hear voices, while there is no auditory accompaniment to visual eye movements. The strength of a voxel's preference for visual mouth movements was strongly correlated with the magnitude of its auditory speech response and its preference for vocal sounds, suggesting that visual and auditory speech features are coded together in small populations of neurons within the pSTS. SIGNIFICANCE STATEMENT Humans interacting face to face make use of auditory cues from the talker's voice and visual cues from the talker's mouth to understand speech. The human posterior superior temporal sulcus (pSTS), a brain region known to be important for speech perception, is complex, with some regions responding to specific visual stimuli and others to specific auditory stimuli. Using BOLD fMRI, we show that the natural statistics of human speech, in which voices co-occur with mouth movements, are reflected in the neural architecture of the pSTS. Different pSTS regions prefer visually presented faces containing either a moving mouth or moving eyes, but only mouth-preferring regions respond strongly to voices. PMID:28179553

  13. Mouth and Voice: A Relationship between Visual and Auditory Preference in the Human Superior Temporal Sulcus.

    PubMed

    Zhu, Lin L; Beauchamp, Michael S

    2017-03-08

    Cortex in and around the human posterior superior temporal sulcus (pSTS) is known to be critical for speech perception. The pSTS responds to both the visual modality (especially biological motion) and the auditory modality (especially human voices). Using fMRI in single subjects with no spatial smoothing, we show that visual and auditory selectivity are linked. Regions of the pSTS were identified that preferred visually presented moving mouths (presented in isolation or as part of a whole face) or moving eyes. Mouth-preferring regions responded strongly to voices and showed a significant preference for vocal compared with nonvocal sounds. In contrast, eye-preferring regions did not respond to either vocal or nonvocal sounds. The converse was also true: regions of the pSTS that showed a significant response to speech or preferred vocal to nonvocal sounds responded more strongly to visually presented mouths than eyes. These findings can be explained by environmental statistics. In natural environments, humans see visual mouth movements at the same time as they hear voices, while there is no auditory accompaniment to visual eye movements. The strength of a voxel's preference for visual mouth movements was strongly correlated with the magnitude of its auditory speech response and its preference for vocal sounds, suggesting that visual and auditory speech features are coded together in small populations of neurons within the pSTS. SIGNIFICANCE STATEMENT Humans interacting face to face make use of auditory cues from the talker's voice and visual cues from the talker's mouth to understand speech. The human posterior superior temporal sulcus (pSTS), a brain region known to be important for speech perception, is complex, with some regions responding to specific visual stimuli and others to specific auditory stimuli. Using BOLD fMRI, we show that the natural statistics of human speech, in which voices co-occur with mouth movements, are reflected in the neural architecture of the pSTS. Different pSTS regions prefer visually presented faces containing either a moving mouth or moving eyes, but only mouth-preferring regions respond strongly to voices. Copyright © 2017 the authors 0270-6474/17/372697-12$15.00/0.

  14. Ultrasound Algorithm Derivation for Soil Moisture Content Estimation

    NASA Technical Reports Server (NTRS)

    Belisle, W.R.; Metzl, R.; Choi, J.; Aggarwal, M. D.; Coleman, T.

    1997-01-01

    Soil moisture content can be estimated by evaluating the velocity at which sound waves travel through a known volume of solid material. This research involved the development of three soil algorithms relating the moisture content to the velocity at which sound waves moved through dry and moist media. Pressure and shear wave propagation equations were used in conjunction with soil property descriptions to derive algorithms appropriate for describing the effects of moisture content variation on the velocity of sound waves in soils with and without complete soil pore water volumes, An elementary algorithm was used to estimate soil moisture contents ranging from 0.08 g/g to 0.5 g/g from sound wave velocities ranging from 526 m/s to 664 m/s. Secondary algorithms were also used to estimate soil moisture content from sound wave velocities through soils with pores that were filled predominantly with air or water.

  15. STS-57 Earth observation of King Sound in northwest Australia

    NASA Technical Reports Server (NTRS)

    1993-01-01

    STS-57 Earth observation taken aboard Endeavour, Orbiter Vehicle (OV) 105, is of King Sound in northwest Australia. Roebuck Bay with the city of Broom on its northern shore is south of King Sound. Sediment in the sound is deposited by the Fitzroy River, which is the major body draining the Kimberley Plateau about 200 miles to the west. The extent of the tidal flats around the Sound is indicated by the large white areas covered with a salty residue. According to NASA scientists studying the STS-57 Earth photos, northwest wind gusts are ruffling areas of the water's surface at the mouth of King Sound and in neighboring Collier Bay. Therefore the water is less reflective and dark. The higher reflectance on the brightest areas is caused by biological oils floating on the surface and reducing the capillary wave action. The scientists point out that the oils take the forms of the currents and eddies in the picture. These eddies indicate that the water offshore is moving at a different speed

  16. Sound field reproduction as an equivalent acoustical scattering problem.

    PubMed

    Fazi, Filippo Maria; Nelson, Philip A

    2013-11-01

    Given a continuous distribution of acoustic sources, the determination of the source strength that ensures the synthesis of a desired sound field is shown to be identical to the solution of an equivalent acoustic scattering problem. The paper begins with the presentation of the general theory that underpins sound field reproduction with secondary sources continuously arranged on the boundary of the reproduction region. The process of reproduction by a continuous source distribution is modeled by means of an integral operator (the single layer potential). It is then shown how the solution of the sound reproduction problem corresponds to that of an equivalent scattering problem. Analytical solutions are computed for two specific instances of this problem, involving, respectively, the use of a secondary source distribution in spherical and planar geometries. The results are shown to be the same as those obtained with analyses based on High Order Ambisonics and Wave Field Synthesis, respectively, thus bringing to light a fundamental analogy between these two methods of sound reproduction. Finally, it is shown how the physical optics (Kirchhoff) approximation enables the derivation of a high-frequency simplification for the problem under consideration, this in turn being related to the secondary source selection criterion reported in the literature on Wave Field Synthesis.

  17. Investigation of spherical loudspeaker arrays for local active control of sound.

    PubMed

    Peleg, Tomer; Rafaely, Boaz

    2011-10-01

    Active control of sound can be employed globally to reduce noise levels in an entire enclosure, or locally around a listener's head. Recently, spherical loudspeaker arrays have been studied as multiple-channel sources for local active control of sound, presenting the fundamental theory and several active control configurations. In this paper, important aspects of using a spherical loudspeaker array for local active control of sound are further investigated. First, the feasibility of creating sphere-shaped quiet zones away from the source is studied both theoretically and numerically, showing that these quiet zones are associated with sound amplification and poor system robustness. To mitigate the latter, the design of shell-shaped quiet zones around the source is investigated. A combination of two spherical sources is then studied with the aim of enlarging the quiet zone. The two sources are employed to generate quiet zones that surround a rigid sphere, investigating the application of active control around a listener's head. A significant improvement in performance is demonstrated in this case over a conventional headrest-type system that uses two monopole secondary sources. Finally, several simulations are presented to support the theoretical work and to demonstrate the performance and limitations of the system. © 2011 Acoustical Society of America

  18. Echolocation versus echo suppression in humans

    PubMed Central

    Wallmeier, Ludwig; Geßele, Nikodemus; Wiegrebe, Lutz

    2013-01-01

    Several studies have shown that blind humans can gather spatial information through echolocation. However, when localizing sound sources, the precedence effect suppresses spatial information of echoes, and thereby conflicts with effective echolocation. This study investigates the interaction of echolocation and echo suppression in terms of discrimination suppression in virtual acoustic space. In the ‘Listening’ experiment, sighted subjects discriminated between positions of a single sound source, the leading or the lagging of two sources, respectively. In the ‘Echolocation’ experiment, the sources were replaced by reflectors. Here, the same subjects evaluated echoes generated in real time from self-produced vocalizations and thereby discriminated between positions of a single reflector, the leading or the lagging of two reflectors, respectively. Two key results were observed. First, sighted subjects can learn to discriminate positions of reflective surfaces echo-acoustically with accuracy comparable to sound source discrimination. Second, in the Listening experiment, the presence of the leading source affected discrimination of lagging sources much more than vice versa. In the Echolocation experiment, however, the presence of both the lead and the lag strongly affected discrimination. These data show that the classically described asymmetry in the perception of leading and lagging sounds is strongly diminished in an echolocation task. Additional control experiments showed that the effect is owing to both the direct sound of the vocalization that precedes the echoes and owing to the fact that the subjects actively vocalize in the echolocation task. PMID:23986105

  19. Two dimensional sound field reproduction using higher order sources to exploit room reflections.

    PubMed

    Betlehem, Terence; Poletti, Mark A

    2014-04-01

    In this paper, sound field reproduction is performed in a reverberant room using higher order sources (HOSs) and a calibrating microphone array. Previously a sound field was reproduced with fixed directivity sources and the reverberation compensated for using digital filters. However by virtue of their directive properties, HOSs may be driven to not only avoid the creation of excess reverberation but also to use room reflection to contribute constructively to the desired sound field. The manner by which the loudspeakers steer the sound around the room is determined by measuring the acoustic transfer functions. The requirements on the number and order N of HOSs for accurate reproduction in a reverberant room are derived, showing a 2N + 1-fold decrease in the number of loudspeakers in comparison to using monopole sources. HOSs are shown applicable to rooms with a rich variety of wall reflections while in an anechoic room their advantages may be lost. Performance is investigated in a room using extensions of both the diffuse field model and a more rigorous image-source simulation method, which account for the properties of the HOSs. The robustness of the proposed method is validated by introducing measurement errors.

  20. Noise pair velocity and range echo location system

    DOEpatents

    Erskine, D.J.

    1999-02-16

    An echo-location method for microwaves, sound and light capable of using incoherent and arbitrary waveforms of wide bandwidth to measure velocity and range (and target size) simultaneously to high resolution is disclosed. Two interferometers having very long and nearly equal delays are used in series with the target interposed. The delays can be longer than the target range of interest. The first interferometer imprints a partial coherence on an initially incoherent source which allows autocorrelation to be performed on the reflected signal to determine velocity. A coherent cross-correlation subsequent to the second interferometer with the source determines a velocity discriminated range. Dithering the second interferometer identifies portions of the cross-correlation belonging to a target apart from clutter moving at a different velocity. The velocity discrimination is insensitive to all slowly varying distortions in the signal path. Speckle in the image of target and antenna lobing due to parasitic reflections is minimal for an incoherent source. An arbitrary source which varies its spectrum dramatically and randomly from pulse to pulse creates a radar elusive to jamming. Monochromatic sources which jigger in frequency from pulse to pulse or combinations of monochromatic sources can simulate some benefits of incoherent broadband sources. Clutter which has a symmetrical velocity spectrum will self-cancel for short wavelengths, such as the apparent motion of ground surrounding target from a sidelooking airborne antenna. 46 figs.

  1. Noise pair velocity and range echo location system

    DOEpatents

    Erskine, David J.

    1999-01-01

    An echo-location method for microwaves, sound and light capable of using incoherent and arbitrary waveforms of wide bandwidth to measure velocity and range (and target size) simultaneously to high resolution. Two interferometers having very long and nearly equal delays are used in series with the target interposed. The delays can be longer than the target range of interest. The first interferometer imprints a partial coherence on an initially incoherent source which allows autocorrelation to be performed on the reflected signal to determine velocity. A coherent cross-correlation subsequent to the second interferometer with the source determines a velocity discriminated range. Dithering the second interferometer identifies portions of the cross-correlation belonging to a target apart from clutter moving at a different velocity. The velocity discrimination is insensitive to all slowly varying distortions in the signal path. Speckle in the image of target and antenna lobing due to parasitic reflections is minimal for an incoherent source. An arbitrary source which varies its spectrum dramatically and randomly from pulse to pulse creates a radar elusive to jamming. Monochromatic sources which jigger in frequency from pulse to pulse or combinations of monochromatic sources can simulate some benefits of incoherent broadband sources. Clutter which has a symmetrical velocity spectrum will self-cancel for short wavelengths, such as the apparent motion of ground surrounding target from a sidelooking airborne antenna.

  2. Pectoral sound generation in the blue catfish Ictalurus furcatus.

    PubMed

    Mohajer, Yasha; Ghahramani, Zachary; Fine, Michael L

    2015-03-01

    Catfishes produce pectoral stridulatory sounds by "jerk" movements that rub ridges on the dorsal process against the cleithrum. We recorded sound synchronized with high-speed video to investigate the hypothesis that blue catfish Ictalurus furcatus produce sounds by a slip-stick mechanism, previously described only in invertebrates. Blue catfish produce a variably paced series of sound pulses during abduction sweeps (pulsers) although some individuals (sliders) form longer duration sound units (slides) interspersed with pulses. Typical pulser sounds are evoked by short 1-2 ms movements with a rotation of 2°-3°. Jerks excite sounds that increase in amplitude after motion stops, suggesting constructive interference, which decays before the next jerk. Longer contact of the ridges produces a more steady-state sound in slides. Pulse pattern during stridulation is determined by pauses without movement: the spine moves during about 14 % of the abduction sweep in pulsers (~45 % in sliders) although movement appears continuous to the human eye. Spine rotation parameters do not predict pulse amplitude, but amplitude correlates with pause duration suggesting that force between the dorsal process and cleithrum increases with longer pauses. Sound production, stimulated by a series of rapid movements that set the pectoral girdle into resonance, is caused by a slip-stick mechanism.

  3. Cognitive load of navigating without vision when guided by virtual sound versus spatial language.

    PubMed

    Klatzky, Roberta L; Marston, James R; Giudice, Nicholas A; Golledge, Reginald G; Loomis, Jack M

    2006-12-01

    A vibrotactile N-back task was used to generate cognitive load while participants were guided along virtual paths without vision. As participants stepped in place, they moved along a virtual path of linear segments. Information was provided en route about the direction of the next turning point, by spatial language ("left," "right," or "straight") or virtual sound (i.e., the perceived azimuth of the sound indicated the target direction). The authors hypothesized that virtual sound, being processed at direct perceptual levels, would have lower load than even simple language commands, which require cognitive mediation. As predicted, whereas the guidance modes did not differ significantly in the no-load condition, participants showed shorter distance traveled and less time to complete a path when performing the N-back task while navigating with virtual sound as guidance. Virtual sound also produced better N-back performance than spatial language. By indicating the superiority of virtual sound for guidance when cognitive load is present, as is characteristic of everyday navigation, these results have implications for guidance systems for the visually impaired and others.

  4. Seismic and Biological Sources of Ambient Ocean Sound

    NASA Astrophysics Data System (ADS)

    Freeman, Simon Eric

    Sound is the most efficient radiation in the ocean. Sounds of seismic and biological origin contain information regarding the underlying processes that created them. A single hydrophone records summary time-frequency information from the volume within acoustic range. Beamforming using a hydrophone array additionally produces azimuthal estimates of sound sources. A two-dimensional array and acoustic focusing produce an unambiguous two-dimensional `image' of sources. This dissertation describes the application of these techniques in three cases. The first utilizes hydrophone arrays to investigate T-phases (water-borne seismic waves) in the Philippine Sea. Ninety T-phases were recorded over a 12-day period, implying a greater number of seismic events occur than are detected by terrestrial seismic monitoring in the region. Observation of an azimuthally migrating T-phase suggests that reverberation of such sounds from bathymetric features can occur over megameter scales. In the second case, single hydrophone recordings from coral reefs in the Line Islands archipelago reveal that local ambient reef sound is spectrally similar to sounds produced by small, hard-shelled benthic invertebrates in captivity. Time-lapse photography of the reef reveals an increase in benthic invertebrate activity at sundown, consistent with an increase in sound level. The dominant acoustic phenomenon on these reefs may thus originate from the interaction between a large number of small invertebrates and the substrate. Such sounds could be used to take census of hard-shelled benthic invertebrates that are otherwise extremely difficult to survey. A two-dimensional `map' of sound production over a coral reef in the Hawaiian Islands was obtained using two-dimensional hydrophone array in the third case. Heterogeneously distributed bio-acoustic sources were generally co-located with rocky reef areas. Acoustically dominant snapping shrimp were largely restricted to one location within the area surveyed. This distribution of sources could reveal small-scale spatial ecological limitations, such as the availability of food and shelter. While array-based passive acoustic sensing is well established in seismoacoustics, the technique is little utilized in the study of ambient biological sound. With the continuance of Moore's law and advances in battery and memory technology, inferring biological processes from ambient sound may become a more accessible tool in underwater ecological evaluation and monitoring.

  5. Calculating far-field radiated sound pressure levels from NASTRAN output

    NASA Technical Reports Server (NTRS)

    Lipman, R. R.

    1986-01-01

    FAFRAP is a computer program which calculates far field radiated sound pressure levels from quantities computed by a NASTRAN direct frequency response analysis of an arbitrarily shaped structure. Fluid loading on the structure can be computed directly by NASTRAN or an added-mass approximation to fluid loading on the structure can be used. Output from FAFRAP includes tables of radiated sound pressure levels and several types of graphic output. FAFRAP results for monopole and dipole sources compare closely with an explicit calculation of the radiated sound pressure level for those sources.

  6. Modelling of human low frequency sound localization acuity demonstrates dominance of spatial variation of interaural time difference and suggests uniform just-noticeable differences in interaural time difference.

    PubMed

    Smith, Rosanna C G; Price, Stephen R

    2014-01-01

    Sound source localization is critical to animal survival and for identification of auditory objects. We investigated the acuity with which humans localize low frequency, pure tone sounds using timing differences between the ears. These small differences in time, known as interaural time differences or ITDs, are identified in a manner that allows localization acuity of around 1° at the midline. Acuity, a relative measure of localization ability, displays a non-linear variation as sound sources are positioned more laterally. All species studied localize sounds best at the midline and progressively worse as the sound is located out towards the side. To understand why sound localization displays this variation with azimuthal angle, we took a first-principles, systemic, analytical approach to model localization acuity. We calculated how ITDs vary with sound frequency, head size and sound source location for humans. This allowed us to model ITD variation for previously published experimental acuity data and determine the distribution of just-noticeable differences in ITD. Our results suggest that the best-fit model is one whereby just-noticeable differences in ITDs are identified with uniform or close to uniform sensitivity across the physiological range. We discuss how our results have several implications for neural ITD processing in different species as well as development of the auditory system.

  7. Graphene-on-paper sound source devices.

    PubMed

    Tian, He; Ren, Tian-Ling; Xie, Dan; Wang, Yu-Feng; Zhou, Chang-Jian; Feng, Ting-Ting; Fu, Di; Yang, Yi; Peng, Ping-Gang; Wang, Li-Gang; Liu, Li-Tian

    2011-06-28

    We demonstrate an interesting phenomenon that graphene can emit sound. The application of graphene can be expanded in the acoustic field. Graphene-on-paper sound source devices are made by patterning graphene on paper substrates. Three graphene sheet samples with the thickness of 100, 60, and 20 nm were fabricated. Sound emission from graphene is measured as a function of power, distance, angle, and frequency in the far-field. The theoretical model of air/graphene/paper/PCB board multilayer structure is established to analyze the sound directivity, frequency response, and efficiency. Measured sound pressure level (SPL) and efficiency are in good agreement with theoretical results. It is found that graphene has a significant flat frequency response in the wide ultrasound range 20-50 kHz. In addition, the thinner graphene sheets can produce higher SPL due to its lower heat capacity per unit area (HCPUA). The infrared thermal images reveal that a thermoacoustic effect is the working principle. We find that the sound performance mainly depends on the HCPUA of the conductor and the thermal properties of the substrate. The paper-based graphene sound source devices have highly reliable, flexible, no mechanical vibration, simple structure and high performance characteristics. It could open wide applications in multimedia, consumer electronics, biological, medical, and many other areas.

  8. Representation of Sound Objects within Early-Stage Auditory Areas: A Repetition Effect Study Using 7T fMRI

    PubMed Central

    Da Costa, Sandra; Bourquin, Nathalie M.-P.; Knebel, Jean-François; Saenz, Melissa; van der Zwaag, Wietske; Clarke, Stephanie

    2015-01-01

    Environmental sounds are highly complex stimuli whose recognition depends on the interaction of top-down and bottom-up processes in the brain. Their semantic representations were shown to yield repetition suppression effects, i. e. a decrease in activity during exposure to a sound that is perceived as belonging to the same source as a preceding sound. Making use of the high spatial resolution of 7T fMRI we have investigated the representations of sound objects within early-stage auditory areas on the supratemporal plane. The primary auditory cortex was identified by means of tonotopic mapping and the non-primary areas by comparison with previous histological studies. Repeated presentations of different exemplars of the same sound source, as compared to the presentation of different sound sources, yielded significant repetition suppression effects within a subset of early-stage areas. This effect was found within the right hemisphere in primary areas A1 and R as well as two non-primary areas on the antero-medial part of the planum temporale, and within the left hemisphere in A1 and a non-primary area on the medial part of Heschl’s gyrus. Thus, several, but not all early-stage auditory areas encode the meaning of environmental sounds. PMID:25938430

  9. Conversion of environmental data to a digital-spatial database, Puget Sound area, Washington

    USGS Publications Warehouse

    Uhrich, M.A.; McGrath, T.S.

    1997-01-01

    Data and maps from the Puget Sound Environmental Atlas, compiled for the U.S. Environmental Protection Agency, the Puget Sound Water Quality Authority, and the U.S. Army Corps of Engineers, have been converted into a digital-spatial database using a geographic information system. Environmental data for the Puget Sound area,collected from sources other than the Puget SoundEnvironmental Atlas by different Federal, State, andlocal agencies, also have been converted into thisdigital-spatial database. Background on the geographic-information-system planning process, the design and implementation of the geographic information-system database, and the reasons for conversion to this digital-spatial database are included in this report. The Puget Sound Environmental Atlas data layers include information about seabird nesting areas, eelgrass and kelp habitat, marine mammal and fish areas, and shellfish resources and bed certification. Data layers, from sources other than the Puget Sound Environmental Atlas, include the Puget Sound shoreline, the water-body system, shellfish growing areas, recreational shellfish beaches, sewage-treatment outfalls, upland hydrography,watershed and political boundaries, and geographicnames. The sources of data, descriptions of the datalayers, and the steps and errors of processing associated with conversion to a digital-spatial database used in development of the Puget Sound Geographic Information System also are included in this report. The appendixes contain data dictionaries for each of the resource layers and error values for the conversion of Puget SoundEnvironmental Atlas data.

  10. Further Progress in Noise Source Identification in High Speed Jets via Causality Principle

    NASA Technical Reports Server (NTRS)

    Panda, J.; Seasholtz, R. G.; Elam, K. A.

    2004-01-01

    To locate noise sources in high-speed jets, the sound pressure fluctuations p/, measured at far field locations, were correlated with each of density p, axial velocity u, radial velocity v, puu and pvv fluctuations measured from various points in fully expanded, unheated plumes of Mach number 0.95, 1.4 and 1.8. The velocity and density fluctuations were measured simultaneously using a recently developed, non-intrusive, point measurement technique based on molecular Rayleigh scattering (Seasholtz, Panda, and Elam, AIAA Paper 2002-0827). The technique uses a continuous wave, narrow line-width laser, Fabry-Perot interferometer and photon counting electronics. The far field sound pressure fluctuations at 30 to the jet axis provided the highest correlation coefficients with all flow variables. The correlation coefficients decreased sharply with increased microphone polar angle, and beyond about 60 all correlation mostly fell below the experimental noise floor. Among all correlations < puu; p/> showed the highest values. Interestingly, , in all respects, were very similar to . The and correlations with 90deg microphone fell below the noise floor. By moving the laser probe at various locations in the jet it was found that the strongest noise source lies downstream of the end of the potential core and extends many diameters beyond. Correlation measurement from the lip shear layer showed a Mach number dependency. While significant correlations were measured in Mach 1.8 jet, values were mostly below the noise floor for subsonic Mach 0.95 jet. Various additional analyses showed that fluctuations from large organized structures mostly contributed to the measured correlation, while that from small scale structures fell below the noise floor.

  11. Sound localization skills in children who use bilateral cochlear implants and in children with normal acoustic hearing

    PubMed Central

    Grieco-Calub, Tina M.; Litovsky, Ruth Y.

    2010-01-01

    Objectives To measure sound source localization in children who have sequential bilateral cochlear implants (BICIs); to determine if localization accuracy correlates with performance on a right-left discrimination task (i.e., spatial acuity); to determine if there is a measurable bilateral benefit on a sound source identification task (i.e., localization accuracy) by comparing performance under bilateral and unilateral listening conditions; to determine if sound source localization continues to improve with longer durations of bilateral experience. Design Two groups of children participated in this study: a group of 21 children who received BICIs in sequential procedures (5–14 years old) and a group of 7 typically-developing children with normal acoustic hearing (5 years old). Testing was conducted in a large sound-treated booth with loudspeakers positioned on a horizontal arc with a radius of 1.2 m. Children participated in two experiments that assessed spatial hearing skills. Spatial hearing acuity was assessed with a discrimination task in which listeners determined if a sound source was presented on the right or left side of center; the smallest angle at which performance on this task was reliably above chance is the minimum audible angle. Sound localization accuracy was assessed with a sound source identification task in which children identified the perceived position of the sound source from a multi-loudspeaker array (7 or 15); errors are quantified using the root-mean-square (RMS) error. Results Sound localization accuracy was highly variable among the children with BICIs, with RMS errors ranging from 19°–56°. Performance of the NH group, with RMS errors ranging from 9°–29° was significantly better. Within the BICI group, in 11/21 children RMS errors were smaller in the bilateral vs. unilateral listening condition, indicating bilateral benefit. There was a significant correlation between spatial acuity and sound localization accuracy (R2=0.68, p<0.01), suggesting that children who achieve small RMS errors tend to have the smallest MAAs. Although there was large intersubject variability, testing of 11 children in the BICI group at two sequential visits revealed a subset of children who show improvement in spatial hearing skills over time. Conclusions A subset of children who use sequential BICIs can acquire sound localization abilities, even after long intervals between activation of hearing in the first- and second-implanted ears. This suggests that children with activation of the second implant later in life may be capable of developing spatial hearing abilities. The large variability in performance among the children with BICIs suggests that maturation of sound localization abilities in children with BICIs may be dependent on various individual subject factors such as age of implantation and chronological age. PMID:20592615

  12. Sound Waves Levitate Substrates

    NASA Technical Reports Server (NTRS)

    Lee, M. C.; Wang, T. G.

    1982-01-01

    System recently tested uses acoustic waves to levitate liquid drops, millimeter-sized glass microballoons, and other objects for coating by vapor deposition or capillary attraction. Cylindrical contactless coating/handling facility employs a cylindrical acoustic focusing radiator and a tapered reflector to generate a specially-shaped standing wave pattern. Article to be processed is captured by the acoustic force field under the reflector and moves as reflector is moved to different work stations.

  13. 15. Detail of 1946 chain rack alot for moving blocks ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    15. Detail of 1946 chain rack alot for moving blocks on floor of drydock in far SE corner of South Section. In 1994 theme chain rack alots and accompanying blocking mechanisms were no longer in use and were extremely deteriorated. For view of system as originally installed, see 1946 historic copy photograph WA-116-E-33. - Puget Sound Naval Shipyard, Drydock No. 3, Farragut Avenue, Bremerton, Kitsap County, WA

  14. Hardwall acoustical characteristics and measurement capabilities of the NASA Lewis 9 x 15 foot low speed wind tunnel

    NASA Technical Reports Server (NTRS)

    Rentz, P. E.

    1976-01-01

    Experimental evaluations of the acoustical characteristics and source sound power and directionality measurement capabilities of the NASA Lewis 9 x 15 foot low speed wind tunnel in the untreated or hardwall configuration were performed. The results indicate that source sound power estimates can be made using only settling chamber sound pressure measurements. The accuracy of these estimates, expressed as one standard deviation, can be improved from + or - 4 db to + or - 1 db if sound pressure measurements in the preparation room and diffuser are also used and source directivity information is utilized. A simple procedure is presented. Acceptably accurate measurements of source direct field acoustic radiation were found to be limited by the test section reverberant characteristics to 3.0 feet for omni-directional and highly directional sources. Wind-on noise measurements in the test section, settling chamber and preparation room were found to depend on the sixth power of tunnel velocity. The levels were compared with various analytic models. Results are presented and discussed.

  15. Reduced order modeling of head related transfer functions for virtual acoustic displays

    NASA Astrophysics Data System (ADS)

    Willhite, Joel A.; Frampton, Kenneth D.; Grantham, D. Wesley

    2003-04-01

    The purpose of this work is to improve the computational efficiency in acoustic virtual applications by creating and testing reduced order models of the head related transfer functions used in localizing sound sources. State space models of varying order were generated from zero-elevation Head Related Impulse Responses (HRIRs) using Kungs Single Value Decomposition (SVD) technique. The inputs to the models are the desired azimuths of the virtual sound sources (from minus 90 deg to plus 90 deg, in 10 deg increments) and the outputs are the left and right ear impulse responses. Trials were conducted in an anechoic chamber in which subjects were exposed to real sounds that were emitted by individual speakers across a numbered speaker array, phantom sources generated from the original HRIRs, and phantom sound sources generated with the different reduced order state space models. The error in the perceived direction of the phantom sources generated from the reduced order models was compared to errors in localization using the original HRIRs.

  16. Estimation of multiple sound sources with data and model uncertainties using the EM and evidential EM algorithms

    NASA Astrophysics Data System (ADS)

    Wang, Xun; Quost, Benjamin; Chazot, Jean-Daniel; Antoni, Jérôme

    2016-01-01

    This paper considers the problem of identifying multiple sound sources from acoustical measurements obtained by an array of microphones. The problem is solved via maximum likelihood. In particular, an expectation-maximization (EM) approach is used to estimate the sound source locations and strengths, the pressure measured by a microphone being interpreted as a mixture of latent signals emitted by the sources. This work also considers two kinds of uncertainties pervading the sound propagation and measurement process: uncertain microphone locations and uncertain wavenumber. These uncertainties are transposed to the data in the belief functions framework. Then, the source locations and strengths can be estimated using a variant of the EM algorithm, known as the Evidential EM (E2M) algorithm. Eventually, both simulation and real experiments are shown to illustrate the advantage of using the EM in the case without uncertainty and the E2M in the case of uncertain measurement.

  17. Auditory Localization: An Annotated Bibliography

    DTIC Science & Technology

    1983-11-01

    tranverse plane, natural sound localization in ,-- both horizontal and vertical planes can be performed with nearly the same accuracy as real sound sources...important for unscrambling the competing sounds which so often occur in natural environments. A workable sound sensor has been constructed and empirical

  18. Heart failure - tests

    MedlinePlus

    CHF - tests; Congestive heart failure - tests; Cardiomyopathy - tests; HF - tests ... An echocardiogram (echo) is a test that uses sound waves to create a moving picture of the heart. The picture is much more detailed than a plain ...

  19. Air Reactions to Objects Moving at Rates Above the Velocity of Sound with Application to the Air Propeller

    NASA Technical Reports Server (NTRS)

    Reed, S Albert

    1922-01-01

    There has been a tradition general among aeronautical engineers that a critical point exists for tip speeds at or near the velocity of sound, indicating a physical limit in the use of propellers at higher tip speeds; the idea being that something would occur analogous to what is known in marine propellers as cavitation. In the examination of the physics pertaining to both propellers and projectiles moving at or above 1100 feet per second, the conclusion was reached by the author that there is no reason for the existence of such a critical point and that, if it had been noted by observers it was not inherent in the phenomena revealed, but rather due to a particular shape or proportion of the projectile and that, with properly proportioned sections, it would not exist.

  20. Rotorcraft Noise Model

    NASA Technical Reports Server (NTRS)

    Lucas, Michael J.; Marcolini, Michael A.

    1997-01-01

    The Rotorcraft Noise Model (RNM) is an aircraft noise impact modeling computer program being developed for NASA-Langley Research Center which calculates sound levels at receiver positions either on a uniform grid or at specific defined locations. The basic computational model calculates a variety of metria. Acoustic properties of the noise source are defined by two sets of sound pressure hemispheres, each hemisphere being centered on a noise source of the aircraft. One set of sound hemispheres provides the broadband data in the form of one-third octave band sound levels. The other set of sound hemispheres provides narrowband data in the form of pure-tone sound pressure levels and phase. Noise contours on the ground are output graphically or in tabular format, and are suitable for inclusion in Environmental Impact Statements or Environmental Assessments.

  1. Timing matters: sonar call groups facilitate target localization in bats.

    PubMed

    Kothari, Ninad B; Wohlgemuth, Melville J; Hulgard, Katrine; Surlykke, Annemarie; Moss, Cynthia F

    2014-01-01

    To successfully negotiate a cluttered environment, an echolocating bat must control the timing of motor behaviors in response to dynamic sensory information. Here we detail the big brown bat's adaptive temporal control over sonar call production for tracking prey, moving predictably or unpredictably, under different experimental conditions. We studied the adaptive control of vocal-motor behaviors in free-flying big brown bats, Eptesicus fuscus, as they captured tethered and free-flying insects, in open and cluttered environments. We also studied adaptive sonar behavior in bats trained to track moving targets from a resting position. In each of these experiments, bats adjusted the features of their calls to separate target and clutter. Under many task conditions, flying bats produced prominent sonar sound groups identified as clusters of echolocation pulses with relatively stable intervals, surrounded by longer pulse intervals. In experiments where bats tracked approaching targets from a resting position, bats also produced sonar sound groups, and the prevalence of these sonar sound groups increased when motion of the target was unpredictable. We hypothesize that sonar sound groups produced during flight, and the sonar call doublets produced by a bat tracking a target from a resting position, help the animal resolve dynamic target location and represent the echo scene in greater detail. Collectively, our data reveal adaptive temporal control over sonar call production that allows the bat to negotiate a complex and dynamic environment.

  2. Timing matters: sonar call groups facilitate target localization in bats

    PubMed Central

    Kothari, Ninad B.; Wohlgemuth, Melville J.; Hulgard, Katrine; Surlykke, Annemarie; Moss, Cynthia F.

    2014-01-01

    To successfully negotiate a cluttered environment, an echolocating bat must control the timing of motor behaviors in response to dynamic sensory information. Here we detail the big brown bat's adaptive temporal control over sonar call production for tracking prey, moving predictably or unpredictably, under different experimental conditions. We studied the adaptive control of vocal-motor behaviors in free-flying big brown bats, Eptesicus fuscus, as they captured tethered and free-flying insects, in open and cluttered environments. We also studied adaptive sonar behavior in bats trained to track moving targets from a resting position. In each of these experiments, bats adjusted the features of their calls to separate target and clutter. Under many task conditions, flying bats produced prominent sonar sound groups identified as clusters of echolocation pulses with relatively stable intervals, surrounded by longer pulse intervals. In experiments where bats tracked approaching targets from a resting position, bats also produced sonar sound groups, and the prevalence of these sonar sound groups increased when motion of the target was unpredictable. We hypothesize that sonar sound groups produced during flight, and the sonar call doublets produced by a bat tracking a target from a resting position, help the animal resolve dynamic target location and represent the echo scene in greater detail. Collectively, our data reveal adaptive temporal control over sonar call production that allows the bat to negotiate a complex and dynamic environment. PMID:24860509

  3. Differential presence of anthropogenic compounds dissolved in the marine waters of Puget Sound, WA and Barkley Sound, BC.

    PubMed

    Keil, Richard; Salemme, Keri; Forrest, Brittany; Neibauer, Jaqui; Logsdon, Miles

    2011-11-01

    Organic compounds were evaluated in March 2010 at 22 stations in Barkley Sound, Vancouver Island Canada and at 66 locations in Puget Sound. Of 37 compounds, 15 were xenobiotics, 8 were determined to have an anthropogenic imprint over natural sources, and 13 were presumed to be of natural or mixed origin. The three most frequently detected compounds were salicyclic acid, vanillin and thymol. The three most abundant compounds were diethylhexyl phthalate (DEHP), ethyl vanillin and benzaldehyde (∼600 n g L(-1) on average). Concentrations of xenobiotics were 10-100 times higher in Puget Sound relative to Barkley Sound. Three compound couplets are used to illustrate the influence of human activity on marine waters; vanillin and ethyl vanillin, salicylic acid and acetylsalicylic acid, and cinnamaldehyde and cinnamic acid. Ratios indicate that anthropogenic activities are the predominant source of these chemicals in Puget Sound. Published by Elsevier Ltd.

  4. A SOUND SOURCE LOCALIZATION TECHNIQUE TO SUPPORT SEARCH AND RESCUE IN LOUD NOISE ENVIRONMENTS

    NASA Astrophysics Data System (ADS)

    Yoshinaga, Hiroshi; Mizutani, Koichi; Wakatsuki, Naoto

    At some sites of earthquakes and other disasters, rescuers search for people buried under rubble by listening for the sounds which they make. Thus developing a technique to localize sound sources amidst loud noise will support such search and rescue operations. In this paper, we discuss an experiment performed to test an array signal processing technique which searches for unperceivable sound in loud noise environments. Two speakers simultaneously played a noise of a generator and a voice decreased by 20 dB (= 1/100 of power) from the generator noise at an outdoor space where cicadas were making noise. The sound signal was received by a horizontally set linear microphone array 1.05 m in length and consisting of 15 microphones. The direction and the distance of the voice were computed and the sound of the voice was extracted and played back as an audible sound by array signal processing.

  5. Mapping the sound field of an erupting submarine volcano using an acoustic glider.

    PubMed

    Matsumoto, Haru; Haxel, Joseph H; Dziak, Robert P; Bohnenstiehl, Delwayne R; Embley, Robert W

    2011-03-01

    An underwater glider with an acoustic data logger flew toward a recently discovered erupting submarine volcano in the northern Lau basin. With the volcano providing a wide-band sound source, recordings from the two-day survey produced a two-dimensional sound level map spanning 1 km (depth) × 40 km(distance). The observed sound field shows depth- and range-dependence, with the first-order spatial pattern being consistent with the predictions of a range-dependent propagation model. The results allow constraining the acoustic source level of the volcanic activity and suggest that the glider provides an effective platform for monitoring natural and anthropogenic ocean sounds. © 2011 Acoustical Society of America

  6. Theoretical and Experimental Aspects of Acoustic Modelling of Engine Exhaust Systems with Applications to a Vacuum Pump

    NASA Astrophysics Data System (ADS)

    Sridhara, Basavapatna Sitaramaiah

    In an internal combustion engine, the engine is the noise source and the exhaust pipe is the main transmitter of noise. Mufflers are often used to reduce engine noise level in the exhaust pipe. To optimize a muffler design, a series of experiments could be conducted using various mufflers installed in the exhaust pipe. For each configuration, the radiated sound pressure could be measured. However, this is not a very efficient method. A second approach would be to develop a scheme involving only a few measurements which can predict the radiated sound pressure at a specified distance from the open end of the exhaust pipe. In this work, the engine exhaust system was modelled as a lumped source-muffler-termination system. An expression for the predicted sound pressure level was derived in terms of the source and termination impedances, and the muffler geometry. The pressure source and monopole radiation models were used for the source and the open end of the exhaust pipe. The four pole parameters were used to relate the acoustic properties at two different cross sections of the muffler and the pipe. The developed formulation was verified through a series of experiments. Two loudspeakers and a reciprocating type vacuum pump were used as sound sources during the tests. The source impedance was measured using the direct, two-load and four-load methods. A simple expansion chamber and a side-branch resonator were used as mufflers. Sound pressure level measurements for the prediction scheme were made for several source-muffler and source-straight pipe combinations. The predicted and measured sound pressure levels were compared for all cases considered. In all cases, correlation of the experimental results and those predicted by the developed expressions was good. Predicted and measured values of the insertion loss of the mufflers were compared. The agreement between the two was good. Also, an error analysis of the four-load method was done.

  7. Ambient Sound-Based Collaborative Localization of Indeterministic Devices

    PubMed Central

    Kamminga, Jacob; Le, Duc; Havinga, Paul

    2016-01-01

    Localization is essential in wireless sensor networks. To our knowledge, no prior work has utilized low-cost devices for collaborative localization based on only ambient sound, without the support of local infrastructure. The reason may be the fact that most low-cost devices are indeterministic and suffer from uncertain input latencies. This uncertainty makes accurate localization challenging. Therefore, we present a collaborative localization algorithm (Cooperative Localization on Android with ambient Sound Sources (CLASS)) that simultaneously localizes the position of indeterministic devices and ambient sound sources without local infrastructure. The CLASS algorithm deals with the uncertainty by splitting the devices into subsets so that outliers can be removed from the time difference of arrival values and localization results. Since Android is indeterministic, we select Android devices to evaluate our approach. The algorithm is evaluated with an outdoor experiment and achieves a mean Root Mean Square Error (RMSE) of 2.18 m with a standard deviation of 0.22 m. Estimated directions towards the sound sources have a mean RMSE of 17.5° and a standard deviation of 2.3°. These results show that it is feasible to simultaneously achieve a relative positioning of both devices and sound sources with sufficient accuracy, even when using non-deterministic devices and platforms, such as Android. PMID:27649176

  8. Amplitude and Wavelength Measurement of Sound Waves in Free Space using a Sound Wave Phase Meter

    NASA Astrophysics Data System (ADS)

    Ham, Sounggil; Lee, Kiwon

    2018-05-01

    We developed a sound wave phase meter (SWPM) and measured the amplitude and wavelength of sound waves in free space. The SWPM consists of two parallel metal plates, where the front plate was operated as a diaphragm. An aluminum perforated plate was additionally installed in front of the diaphragm, and the same signal as that applied to the sound source was applied to the perforated plate. The SWPM measures both the sound wave signal due to the diaphragm vibration and the induction signal due to the electric field of the aluminum perforated plate. Therefore, the two measurement signals interfere with each other due to the phase difference according to the distance between the sound source and the SWPM, and the amplitude of the composite signal that is output as a result is periodically changed. We obtained the wavelength of the sound wave from this periodic amplitude change measured in the free space and compared it with the theoretically calculated values.

  9. Investigation of the Statistics of Pure Tone Sound Power Injection from Low Frequency, Finite Sized Sources in a Reverberant Room

    NASA Technical Reports Server (NTRS)

    Smith, Wayne Farrior

    1973-01-01

    The effect of finite source size on the power statistics in a reverberant room for pure tone excitation was investigated. Theoretical results indicate that the standard deviation of low frequency, pure tone finite sources is always less than that predicted by point source theory and considerably less when the source dimension approaches one-half an acoustic wavelength or greater. A supporting experimental study was conducted utilizing an eight inch loudspeaker and a 30 inch loudspeaker at eleven source positions. The resulting standard deviation of sound power output of the smaller speaker is in excellent agreement with both the derived finite source theory and existing point source theory, if the theoretical data is adjusted to account for experimental incomplete spatial averaging. However, the standard deviation of sound power output of the larger speaker is measurably lower than point source theory indicates, but is in good agreement with the finite source theory.

  10. Localizing nearby sound sources in a classroom: Binaural room impulse responses

    NASA Astrophysics Data System (ADS)

    Shinn-Cunningham, Barbara G.; Kopco, Norbert; Martin, Tara J.

    2005-05-01

    Binaural room impulse responses (BRIRs) were measured in a classroom for sources at different azimuths and distances (up to 1 m) relative to a manikin located in four positions in a classroom. When the listener is far from all walls, reverberant energy distorts signal magnitude and phase independently at each frequency, altering monaural spectral cues, interaural phase differences, and interaural level differences. For the tested conditions, systematic distortion (comb-filtering) from an early intense reflection is only evident when a listener is very close to a wall, and then only in the ear facing the wall. Especially for a nearby source, interaural cues grow less reliable with increasing source laterality and monaural spectral cues are less reliable in the ear farther from the sound source. Reverberation reduces the magnitude of interaural level differences at all frequencies; however, the direct-sound interaural time difference can still be recovered from the BRIRs measured in these experiments. Results suggest that bias and variability in sound localization behavior may vary systematically with listener location in a room as well as source location relative to the listener, even for nearby sources where there is relatively little reverberant energy. .

  11. Localizing nearby sound sources in a classroom: binaural room impulse responses.

    PubMed

    Shinn-Cunningham, Barbara G; Kopco, Norbert; Martin, Tara J

    2005-05-01

    Binaural room impulse responses (BRIRs) were measured in a classroom for sources at different azimuths and distances (up to 1 m) relative to a manikin located in four positions in a classroom. When the listener is far from all walls, reverberant energy distorts signal magnitude and phase independently at each frequency, altering monaural spectral cues, interaural phase differences, and interaural level differences. For the tested conditions, systematic distortion (comb-filtering) from an early intense reflection is only evident when a listener is very close to a wall, and then only in the ear facing the wall. Especially for a nearby source, interaural cues grow less reliable with increasing source laterality and monaural spectral cues are less reliable in the ear farther from the sound source. Reverberation reduces the magnitude of interaural level differences at all frequencies; however, the direct-sound interaural time difference can still be recovered from the BRIRs measured in these experiments. Results suggest that bias and variability in sound localization behavior may vary systematically with listener location in a room as well as source location relative to the listener, even for nearby sources where there is relatively little reverberant energy.

  12. Fetal echocardiography

    MedlinePlus

    ... based gel on your belly. A hand-held probe is moved over the area. The probe sends out sound waves, which bounce off the ... screen. In a transvaginal ultrasound, a much smaller probe is placed into the vagina. A transvaginal ultrasound ...

  13. A precedence effect resolves phantom sound source illusions in the parasitoid fly Ormia ochracea

    PubMed Central

    Lee, Norman; Elias, Damian O.; Mason, Andrew C.

    2009-01-01

    Localizing individual sound sources under reverberant environmental conditions can be a challenge when the original source and its acoustic reflections arrive at the ears simultaneously from different paths that convey ambiguous directional information. The acoustic parasitoid fly Ormia ochracea (Diptera: Tachinidae) relies on a pair of ears exquisitely sensitive to sound direction to localize the 5-kHz tone pulsatile calling song of their host crickets. In nature, flies are expected to encounter a complex sound field with multiple sources and their reflections from acoustic clutter potentially masking temporal information relevant to source recognition and localization. In field experiments, O. ochracea were lured onto a test arena and subjected to small random acoustic asymmetries between 2 simultaneous sources. Most flies successfully localize a single source but some localize a ‘phantom’ source that is a summed effect of both source locations. Such misdirected phonotaxis can be elicited reliably in laboratory experiments that present symmetric acoustic stimulation. By varying onset delay between 2 sources, we test whether hyperacute directional hearing in O. ochracea can function to exploit small time differences to determine source location. Selective localization depends on both the relative timing and location of competing sources. Flies preferred phonotaxis to a forward source. With small onset disparities within a 10-ms temporal window of attention, flies selectively localize the leading source while the lagging source has minimal influence on orientation. These results demonstrate the precedence effect as a mechanism to overcome phantom source illusions that arise from acoustic reflections or competing sources. PMID:19332794

  14. Shock waves and the Ffowcs Williams-Hawkings equation

    NASA Technical Reports Server (NTRS)

    Isom, Morris P.; Yu, Yung H.

    1991-01-01

    The expansion of the double divergence of the generalized Lighthill stress tensor, which is the basis of the concept of the role played by shock and contact discontinuities as sources of dipole and monopole sound, is presently applied to the simplest transonic flows: (1) a fixed wing in steady motion, for which there is no sound field, and (2) a hovering helicopter blade that produces a sound field. Attention is given to the contribution of the shock to sound from the viewpoint of energy conservation; the shock emerges as the source of only the quantity of entropy.

  15. Hearing in three dimensions

    NASA Astrophysics Data System (ADS)

    Shinn-Cunningham, Barbara

    2003-04-01

    One of the key functions of hearing is to help us monitor and orient to events in our environment (including those outside the line of sight). The ability to compute the spatial location of a sound source is also important for detecting, identifying, and understanding the content of a sound source, especially in the presence of competing sources from other positions. Determining the spatial location of a sound source poses difficult computational challenges; however, we perform this complex task with proficiency, even in the presence of noise and reverberation. This tutorial will review the acoustic, psychoacoustic, and physiological processes underlying spatial auditory perception. First, the tutorial will examine how the many different features of the acoustic signals reaching a listener's ears provide cues for source direction and distance, both in anechoic and reverberant space. Then we will discuss psychophysical studies of three-dimensional sound localization in different environments and the basic neural mechanisms by which spatial auditory cues are extracted. Finally, ``virtual reality'' approaches for simulating sounds at different directions and distances under headphones will be reviewed. The tutorial will be structured to appeal to a diverse audience with interests in all fields of acoustics and will incorporate concepts from many areas, such as psychological and physiological acoustics, architectural acoustics, and signal processing.

  16. Active control of sound radiation from a vibrating rectangular panel by sound sources and vibration inputs - An experimental comparison

    NASA Technical Reports Server (NTRS)

    Fuller, C. R.; Hansen, C. H.; Snyder, S. D.

    1991-01-01

    Active control of sound radiation from a rectangular panel by two different methods has been experimentally studied and compared. In the first method a single control force applied directly to the structure is used with a single error microphone located in the radiated acoustic field. Global attenuation of radiated sound was observed to occur by two main mechanisms. For 'on-resonance' excitation, the control force had the effect of increasing the total panel input impedance presented to the nosie source, thus reducing all radiated sound. For 'off-resonance' excitation, the control force tends not significantly to modify the panel total response amplitude but rather to restructure the relative phases of the modes leading to a more complex vibration pattern and a decrease in radiation efficiency. For acoustic control, the second method, the number of acoustic sources required for global reduction was seen to increase with panel modal order. The mechanism in this case was that the acoustic sources tended to create an inverse pressure distribution at the panel surface and thus 'unload' the panel by reducing the panel radiation impedance. In general, control by structural inputs appears more effective than control by acoustic sources for structurally radiated noise.

  17. 8. Elevation view of midsection of west wall of South ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    8. Elevation view of midsection of west wall of South Section. This photo makes an extended elevation panorama with photos WA-116-E-7 and WA-116-E-9. Note that Crane No. 42 also appears in photo WA-116-E-7. This is because the crane moved in the time that the photographer moved the camera. - Puget Sound Naval Shipyard, Drydock No. 3, Farragut Avenue, Bremerton, Kitsap County, WA

  18. Aerofoil broadband and tonal noise modelling using stochastic sound sources and incorporated large scale fluctuations

    NASA Astrophysics Data System (ADS)

    Proskurov, S.; Darbyshire, O. R.; Karabasov, S. A.

    2017-12-01

    The present work discusses modifications to the stochastic Fast Random Particle Mesh (FRPM) method featuring both tonal and broadband noise sources. The technique relies on the combination of incorporated vortex-shedding resolved flow available from Unsteady Reynolds-Averaged Navier-Stokes (URANS) simulation with the fine-scale turbulence FRPM solution generated via the stochastic velocity fluctuations in the context of vortex sound theory. In contrast to the existing literature, our method encompasses a unified treatment for broadband and tonal acoustic noise sources at the source level, thus, accounting for linear source interference as well as possible non-linear source interaction effects. When sound sources are determined, for the sound propagation, Acoustic Perturbation Equations (APE-4) are solved in the time-domain. Results of the method's application for two aerofoil benchmark cases, with both sharp and blunt trailing edges are presented. In each case, the importance of individual linear and non-linear noise sources was investigated. Several new key features related to the unsteady implementation of the method were tested and brought into the equation. Encouraging results have been obtained for benchmark test cases using the new technique which is believed to be potentially applicable to other airframe noise problems where both tonal and broadband parts are important.

  19. Experimental localization of an acoustic sound source in a wind-tunnel flow by using a numerical time-reversal technique.

    PubMed

    Padois, Thomas; Prax, Christian; Valeau, Vincent; Marx, David

    2012-10-01

    The possibility of using the time-reversal technique to localize acoustic sources in a wind-tunnel flow is investigated. While the technique is widespread, it has scarcely been used in aeroacoustics up to now. The proposed method consists of two steps: in a first experimental step, the acoustic pressure fluctuations are recorded over a linear array of microphones; in a second numerical step, the experimental data are time-reversed and used as input data for a numerical code solving the linearized Euler equations. The simulation achieves the back-propagation of the waves from the array to the source and takes into account the effect of the mean flow on sound propagation. The ability of the method to localize a sound source in a typical wind-tunnel flow is first demonstrated using simulated data. A generic experiment is then set up in an anechoic wind tunnel to validate the proposed method with a flow at Mach number 0.11. Monopolar sources are first considered that are either monochromatic or have a narrow or wide-band frequency content. The source position estimation is well-achieved with an error inferior to the wavelength. An application to a dipolar sound source shows that this type of source is also very satisfactorily characterized.

  20. Different categories of living and non-living sound-sources activate distinct cortical networks

    PubMed Central

    Engel, Lauren R.; Frum, Chris; Puce, Aina; Walker, Nathan A.; Lewis, James W.

    2009-01-01

    With regard to hearing perception, it remains unclear as to whether, or the extent to which, different conceptual categories of real-world sounds and related categorical knowledge are differentially represented in the brain. Semantic knowledge representations are reported to include the major divisions of living versus non-living things, plus more specific categories including animals, tools, biological motion, faces, and places—categories typically defined by their characteristic visual features. Here, we used functional magnetic resonance imaging (fMRI) to identify brain regions showing preferential activity to four categories of action sounds, which included non-vocal human and animal actions (living), plus mechanical and environmental sound-producing actions (non-living). The results showed a striking antero-posterior division in cortical representations for sounds produced by living versus non-living sources. Additionally, there were several significant differences by category, depending on whether the task was category-specific (e.g. human or not) versus non-specific (detect end-of-sound). In general, (1) human-produced sounds yielded robust activation in the bilateral posterior superior temporal sulci independent of task. Task demands modulated activation of left-lateralized fronto-parietal regions, bilateral insular cortices, and subcortical regions previously implicated in observation-execution matching, consistent with “embodied” and mirror-neuron network representations subserving recognition. (2) Animal action sounds preferentially activated the bilateral posterior insulae. (3) Mechanical sounds activated the anterior superior temporal gyri and parahippocampal cortices. (4) Environmental sounds preferentially activated dorsal occipital and medial parietal cortices. Overall, this multi-level dissociation of networks for preferentially representing distinct sound-source categories provides novel support for grounded cognition models that may underlie organizational principles for hearing perception. PMID:19465134

  1. The directivity of the sound radiation from panels and openings.

    PubMed

    Davy, John L

    2009-06-01

    This paper presents a method for calculating the directivity of the radiation of sound from a panel or opening, whose vibration is forced by the incidence of sound from the other side. The directivity of the radiation depends on the angular distribution of the incident sound energy in the room or duct in whose wall or end the panel or opening occurs. The angular distribution of the incident sound energy is predicted using a model which depends on the sound absorption coefficient of the room or duct surfaces. If the sound source is situated in the room or duct, the sound absorption coefficient model is used in conjunction with a model for the directivity of the sound source. For angles of radiation approaching 90 degrees to the normal to the panel or opening, the effect of the diffraction by the panel or opening, or by the finite baffle in which the panel or opening is mounted, is included. A simple empirical model is developed to predict the diffraction of sound into the shadow zone when the angle of radiation is greater than 90 degrees to the normal to the panel or opening. The method is compared with published experimental results.

  2. Moving Liquids with Sound: The Physics of Acoustic Droplet Ejection for Robust Laboratory Automation in Life Sciences.

    PubMed

    Hadimioglu, Babur; Stearns, Richard; Ellson, Richard

    2016-02-01

    Liquid handling instruments for life science applications based on droplet formation with focused acoustic energy or acoustic droplet ejection (ADE) were introduced commercially more than a decade ago. While the idea of "moving liquids with sound" was known in the 20th century, the development of precise methods for acoustic dispensing to aliquot life science materials in the laboratory began in earnest in the 21st century with the adaptation of the controlled "drop on demand" acoustic transfer of droplets from high-density microplates for high-throughput screening (HTS) applications. Robust ADE implementations for life science applications achieve excellent accuracy and precision by using acoustics first to sense the liquid characteristics relevant for its transfer, and then to actuate transfer of the liquid with customized application of sound energy to the given well and well fluid in the microplate. This article provides an overview of the physics behind ADE and its central role in both acoustical and rheological aspects of robust implementation of ADE in the life science laboratory and its broad range of ejectable materials. © 2015 Society for Laboratory Automation and Screening.

  3. A Tool for Low Noise Procedures Design and Community Noise Impact Assessment: The Rotorcraft Noise Model (RNM)

    NASA Technical Reports Server (NTRS)

    Conner, David A.; Page, Juliet A.

    2002-01-01

    To improve aircraft noise impact modeling capabilities and to provide a tool to aid in the development of low noise terminal area operations for rotorcraft and tiltrotors, the Rotorcraft Noise Model (RNM) was developed by the NASA Langley Research Center and Wyle Laboratories. RNM is a simulation program that predicts how sound will propagate through the atmosphere and accumulate at receiver locations located on flat ground or varying terrain, for single and multiple vehicle flight operations. At the core of RNM are the vehicle noise sources, input as sound hemispheres. As the vehicle "flies" along its prescribed flight trajectory, the source sound propagation is simulated and accumulated at the receiver locations (single points of interest or multiple grid points) in a systematic time-based manner. These sound signals at the receiver locations may then be analyzed to obtain single event footprints, integrated noise contours, time histories, or numerous other features. RNM may also be used to generate spectral time history data over a ground mesh for the creation of single event sound animation videos. Acoustic properties of the noise source(s) are defined in terms of sound hemispheres that may be obtained from theoretical predictions, wind tunnel experimental results, flight test measurements, or a combination of the three. The sound hemispheres may contain broadband data (source levels as a function of one-third octave band) and pure-tone data (in the form of specific frequency sound pressure levels and phase). A PC executable version of RNM is publicly available and has been adopted by a number of organizations for Environmental Impact Assessment studies of rotorcraft noise. This paper provides a review of the required input data, the theoretical framework of RNM's propagation model and the output results. Code validation results are provided from a NATO helicopter noise flight test as well as a tiltrotor flight test program that used the RNM as a tool to aid in the development of low noise approach profiles.

  4. Popcorn: critical temperature, jump and sound

    PubMed Central

    Virot, Emmanuel; Ponomarenko, Alexandre

    2015-01-01

    Popcorn bursts open, jumps and emits a ‘pop’ sound in some hundredths of a second. The physical origin of these three observations remains unclear in the literature. We show that the critical temperature 180°C at which almost all of popcorn pops is consistent with an elementary pressure vessel scenario. We observe that popcorn jumps with a ‘leg’ of starch which is compressed on the ground. As a result, popcorn is midway between two categories of moving systems: explosive plants using fracture mechanisms and jumping animals using muscles. By synchronizing video recordings with acoustic recordings, we propose that the familiar ‘pop’ sound of the popcorn is caused by the release of water vapour. PMID:25673298

  5. Spatial hearing in Cope’s gray treefrog: I. Open and closed loop experiments on sound localization in the presence and absence of noise

    PubMed Central

    Caldwell, Michael S.; Bee, Mark A.

    2014-01-01

    The ability to reliably locate sound sources is critical to anurans, which navigate acoustically complex breeding choruses when choosing mates. Yet, the factors influencing sound localization performance in frogs remain largely unexplored. We applied two complementary methodologies, open and closed loop playback trials, to identify influences on localization abilities in Cope’s gray treefrog, Hyla chrysoscelis. We examined localization acuity and phonotaxis behavior of females in response to advertisement calls presented from 12 azimuthal angles, at two signal levels, in the presence and absence of noise, and at two noise levels. Orientation responses were consistent with precise localization of sound sources, rather than binary discrimination between sources on either side of the body (lateralization). Frogs were unable to discriminate between sounds arriving from forward and rearward directions, and accurate localization was limited to forward sound presentation angles. Within this region, sound presentation angle had little effect on localization acuity. The presence of noise and low signal-to-noise ratios also did not strongly impair localization ability in open loop trials, but females exhibited reduced phonotaxis performance consistent with impaired localization during closed loop trials. We discuss these results in light of previous work on spatial hearing in anurans. PMID:24504182

  6. The Radiated Field Generated by a Monopole Source in a Short, Rigid, Rectangular Duct. Degree awarded by George Washington Univ.

    NASA Technical Reports Server (NTRS)

    Lakota, Barbara Anne

    1998-01-01

    This thesis develops a method to model the acoustic field generated by a monopole source placed in a moving rectangular duct. The walls of the duct are assumed to be infinitesimally thin and the source is placed at the center of the duct. The total acoustic pressure is written in terms of the free-space pressure, or incident pressure, and the scattered pressure. The scattered pressure is the augmentation to the incident pressure due to the presence of the duct. It satisfies a homogeneous wave equation and is discontinuous across the duct walls. Utilizing an integral representation of the scattered pressure, a set of singular boundary integral equations governing the unknown jump in scattered pressure is derived. This equation is solved by the method of collocation after representing the jump in pressure as a double series of shape functions. The solution obtained is then substituted back into the integral representation to determine the scattered pressure, and the total acoustic pressure at any point in the field. A few examples are included to illustrate the influence of various geometric and kinematic parameters on the radiated sound field.

  7. Sex differences present in auditory looming perception, absent in auditory recession

    NASA Astrophysics Data System (ADS)

    Neuhoff, John G.; Seifritz, Erich

    2005-04-01

    When predicting the arrival time of an approaching sound source, listeners typically exhibit an anticipatory bias that affords a margin of safety in dealing with looming objects. The looming bias has been demonstrated behaviorally in the laboratory and in the field (Neuhoff 1998, 2001), neurally in fMRI studies (Seifritz et al., 2002), and comparatively in non-human primates (Ghazanfar, Neuhoff, and Logothetis, 2002). In the current work, male and female listeners were presented with three-dimensional looming sound sources and asked to press a button when the source was at the point of closest approach. Females exhibited a significantly greater anticipatory bias than males. Next, listeners were presented with sounds that either approached or receded and then stopped at three different terminal distances. Consistent with the time-to-arrival judgments, female terminal distance judgments for looming sources were significantly closer than male judgments. However, there was no difference between male and female terminal distance judgments for receding sounds. Taken together with the converging behavioral, neural, and comparative evidence, the current results illustrate the environmental salience of looming sounds and suggest that the anticipatory bias for auditory looming may have been shaped by evolution to provide a selective advantage in dealing with looming objects.

  8. Understanding auditory distance estimation by humpback whales: a computational approach.

    PubMed

    Mercado, E; Green, S R; Schneider, J N

    2008-02-01

    Ranging, the ability to judge the distance to a sound source, depends on the presence of predictable patterns of attenuation. We measured long-range sound propagation in coastal waters to assess whether humpback whales might use frequency degradation cues to range singing whales. Two types of neural networks, a multi-layer and a single-layer perceptron, were trained to classify recorded sounds by distance traveled based on their frequency content. The multi-layer network successfully classified received sounds, demonstrating that the distorting effects of underwater propagation on frequency content provide sufficient cues to estimate source distance. Normalizing received sounds with respect to ambient noise levels increased the accuracy of distance estimates by single-layer perceptrons, indicating that familiarity with background noise can potentially improve a listening whale's ability to range. To assess whether frequency patterns predictive of source distance were likely to be perceived by whales, recordings were pre-processed using a computational model of the humpback whale's peripheral auditory system. Although signals processed with this model contained less information than the original recordings, neural networks trained with these physiologically based representations estimated source distance more accurately, suggesting that listening whales should be able to range singers using distance-dependent changes in frequency content.

  9. Common humpback whale (Megaptera novaeangliae) sound types for passive acoustic monitoring.

    PubMed

    Stimpert, Alison K; Au, Whitlow W L; Parks, Susan E; Hurst, Thomas; Wiley, David N

    2011-01-01

    Humpback whales (Megaptera novaeangliae) are one of several baleen whale species in the Northwest Atlantic that coexist with vessel traffic and anthropogenic noise. Passive acoustic monitoring strategies can be used in conservation management, but the first step toward understanding the acoustic behavior of a species is a good description of its acoustic repertoire. Digital acoustic tags (DTAGs) were placed on humpback whales in the Stellwagen Bank National Marine Sanctuary to record and describe the non-song sounds being produced in conjunction with foraging activities. Peak frequencies of sounds were generally less than 1 kHz, but ranged as high as 6 kHz, and sounds were generally less than 1 s in duration. Cluster analysis distilled the dataset into eight groups of sounds with similar acoustic properties. The two most stereotyped and distinctive types ("wops" and "grunts") were also identified aurally as candidates for use in passive acoustic monitoring. This identification of two of the most common sound types will be useful for moving forward conservation efforts on this Northwest Atlantic feeding ground.

  10. In vivo imaging and vibration measurement of Guinea pig cochlea

    NASA Astrophysics Data System (ADS)

    Choudhury, Niloy; Chen, Fangyi; Zheng, Jiefu; Nuttall, Alfred L.; Jacques, Steven L.

    2008-02-01

    An optical coherence tomography (OCT) system was built to acquire in vivo, both images and vibration measurements of the organ of Corti of the guinea pig. The organ of Corti was viewed through a ~500-μm diameter hole in the bony wall of the scala tympani of the first cochlear turn. In imaging mode, the image was acquired as reflectance R(x,z). In vibration mode, the basilar membrane (BM) or reticular lamina (RL) was selected based on the image. Under software control, the system would move the scanning mirrors to bring the sensing volume of the measurement to the desired tissue location. To address the gain stability problem of the homodyne OCT system, arising from the system moving in and out of the quadrature point and also to resolve the 180 degree ambiguity in the phase measurement using an interferometer, a vibration calibration method is developed by adding a vibrating source to the reference arm to monitor the operating point of the interferometric system. Amplitude gain and phase of various cochlear membranes was measured for different sound pressure level (SPL) varying from 65dB SPL to 93 dB SPL.

  11. Analysis of EEG Related Saccadic Eye Movement

    NASA Astrophysics Data System (ADS)

    Funase, Arao; Kuno, Yoshiaki; Okuma, Shigeru; Yagi, Tohru

    Our final goal is to establish the model for saccadic eye movement that connects the saccade and the electroencephalogram(EEG). As the first step toward this goal, we recorded and analyzed the saccade-related EEG. In the study recorded in this paper, we tried detecting a certain EEG that is peculiar to the eye movement. In these experiments, each subject was instructed to point their eyes toward visual targets (LEDs) or the direction of the sound sources (buzzers). In the control cases, the EEG was recorded in the case of no eye movemens. As results, in the visual experiments, we found that the potential of EEG changed sharply on the occipital lobe just before eye movement. Furthermore, in the case of the auditory experiments, similar results were observed. In the case of the visual experiments and auditory experiments without eye movement, we could not observed the EEG changed sharply. Moreover, when the subject moved his/her eyes toward a right-side target, a change in EEG potential was found on the right occipital lobe. On the contrary, when the subject moved his/her eyes toward a left-side target, a sharp change in EEG potential was found on the left occipital lobe.

  12. Tutorial on the Psychophysics and Technology of Virtual Acoustic Displays

    NASA Technical Reports Server (NTRS)

    Wenzel, Elizabeth M.; Null, Cynthia (Technical Monitor)

    1998-01-01

    Virtual acoustics, also known as 3-D sound and auralization, is the simulation of the complex acoustic field experienced by a listener within an environment. Going beyond the simple intensity panning of normal stereo techniques, the goal is to process sounds so that they appear to come from particular locations in three-dimensional space. Although loudspeaker systems are being developed, most of the recent work focuses on using headphones for playback and is the outgrowth of earlier analog techniques. For example, in binaural recording, the sound of an orchestra playing classical music is recorded through small mics in the two "ear canals" of an anthropomorphic artificial or "dummy" head placed in the audience of a concert hall. When the recorded piece is played back over headphones, the listener passively experiences the illusion of hearing the violins on the left and the cellos on the right, along with all the associated echoes, resonances, and ambience of the original environment. Current techniques use digital signal processing to synthesize the acoustical properties that people use to localize a sound source in space. Thus, they provide the flexibility of a kind of digital dummy head, allowing a more active experience in which a listener can both design and move around or interact with a simulated acoustic environment in real time. Such simulations are being developed for a variety of application areas including architectural acoustics, advanced human-computer interfaces, telepresence and virtual reality, navigation aids for the visually-impaired, and as a test bed for psychoacoustical investigations of complex spatial cues. The tutorial will review the basic psychoacoustical cues that determine human sound localization and the techniques used to measure these cues as Head-Related Transfer Functions (HRTFs) for the purpose of synthesizing virtual acoustic environments. The only conclusive test of the adequacy of such simulations is an operational one in which the localization of real and synthesized stimuli are directly compared in psychophysical studies. To this end, the results of psychophysical experiments examining the perceptual validity of the synthesis technique will be reviewed and factors that can enhance perceptual accuracy and realism will be discussed. Of particular interest is the relationship between individual differences in HRTFs and in behavior, the role of reverberant cues in reducing the perceptual errors observed with virtual sound sources, and the importance of developing perceptually valid methods of simplifying the synthesis technique. Recent attempts to implement the synthesis technique in real time systems will also be discussed and an attempt made to interpret their quoted system specifications in terms of perceptual performance. Finally, some critical research and technology development issues for the future will be outlined.

  13. Community Response to Multiple Sound Sources: Integrating Acoustic and Contextual Approaches in the Analysis

    PubMed Central

    Lercher, Peter; De Coensel, Bert; Dekonink, Luc; Botteldooren, Dick

    2017-01-01

    Sufficient data refer to the relevant prevalence of sound exposure by mixed traffic sources in many nations. Furthermore, consideration of the potential effects of combined sound exposure is required in legal procedures such as environmental health impact assessments. Nevertheless, current practice still uses single exposure response functions. It is silently assumed that those standard exposure-response curves accommodate also for mixed exposures—although some evidence from experimental and field studies casts doubt on this practice. The ALPNAP-study population (N = 1641) shows sufficient subgroups with combinations of rail-highway, highway-main road and rail-highway-main road sound exposure. In this paper we apply a few suggested approaches of the literature to investigate exposure-response curves and its major determinants in the case of exposure to multiple traffic sources. Highly/moderate annoyance and full scale mean annoyance served as outcome. The results show several limitations of the current approaches. Even facing the inherent methodological limitations (energy equivalent summation of sound, rating of overall annoyance) the consideration of main contextual factors jointly occurring with the sources (such as vibration, air pollution) or coping activities and judgments of the wider area soundscape increases the variance explanation from up to 8% (bivariate), up to 15% (base adjustments) up to 55% (full contextual model). The added predictors vary significantly, depending on the source combination. (e.g., significant vibration effects with main road/railway, not highway). Although no significant interactions were found, the observed additive effects are of public health importance. Especially in the case of a three source exposure situation the overall annoyance is already high at lower levels and the contribution of the acoustic indicators is small compared with the non-acoustic and contextual predictors. Noise mapping needs to go down to levels of 40 dBA,Lden to ensure the protection of quiet areas and prohibit the silent “filling up” of these areas with new sound sources. Eventually, to better predict the annoyance in the exposure range between 40 and 60 dBA and support the protection of quiet areas in city and rural areas in planning sound indicators need to be oriented at the noticeability of sound and consider other traffic related by-products (air quality, vibration, coping strain) in future studies and environmental impact assessments. PMID:28632198

  14. Simulation and testing of a multichannel system for 3D sound localization

    NASA Astrophysics Data System (ADS)

    Matthews, Edward Albert

    Three-dimensional (3D) audio involves the ability to localize sound anywhere in a three-dimensional space. 3D audio can be used to provide the listener with the perception of moving sounds and can provide a realistic listening experience for applications such as gaming, video conferencing, movies, and concerts. The purpose of this research is to simulate and test 3D audio by incorporating auditory localization techniques in a multi-channel speaker system. The objective is to develop an algorithm that can place an audio event in a desired location by calculating and controlling the gain factors of each speaker. A MATLAB simulation displays the location of the speakers and perceived sound, which is verified through experimentation. The scenario in which the listener is not equidistant from each of the speakers is also investigated and simulated. This research is envisioned to lead to a better understanding of human localization of sound, and will contribute to a more realistic listening experience.

  15. Management implications of broadband sound in modulating wild silver carp (Hypophthalmichthys molitrix) behavior

    USGS Publications Warehouse

    Vetter, Brooke J.; Calfee, Robin D.; Mensinger, Allen F.

    2017-01-01

    Invasive silver carp (Hypophthalmichthys molitrix) dominate large regions of the Mississippi River drainage, outcompete native species, and are notorious for their prolific and unusual jumping behavior. High densities of juvenile and adult (~25 kg) carp are known to jump up to 3 m above the water surface in response to moving watercraft. Broadband sound recorded from an outboard motor (100 hp at 32 km/hr) can modulate their behavior in captivity; however, the response of wild silver carp to broadband sound has yet to be determined. In this experiment, broadband sound (0.06–10 kHz) elicited jumping behavior from silver carp in the Spoon River near Havana, IL independent of boat movement, indicating acoustic stimulus alone is sufficient to induce jumping. Furthermore, the number of jumping fish decreased with subsequent sound exposures. Understanding silver carp jumping is not only important from a behavioral standpoint, it is also critical to determine effective techniques for controlling this harmful species, such as herding fish into a net for removal.

  16. Temporal processing deficit leads to impaired multisensory binding in schizophrenia.

    PubMed

    Zvyagintsev, Mikhail; Parisi, Carmen; Mathiak, Klaus

    2017-09-01

    Schizophrenia has been characterised by neurodevelopmental dysconnectivity resulting in cognitive and perceptual dysmetria. Hence patients with schizophrenia may be impaired to detect the temporal relationship between stimuli in different sensory modalities. However, only a few studies described deficit in perception of temporally asynchronous multisensory stimuli in schizophrenia. We examined the perceptual bias and the processing time of synchronous and delayed sounds in the streaming-bouncing illusion in 16 patients with schizophrenia and a matched control group of 18 participants. Equal for patients and controls, the synchronous sound biased the percept of two moving squares towards bouncing as opposed to the more frequent streaming percept in the condition without sound. In healthy controls, a delay of the sound presentation significantly reduced the bias and led to prolonged processing time whereas patients with schizophrenia did not differentiate between this condition and the condition with synchronous sound. Schizophrenia leads to a prolonged window of simultaneity for audiovisual stimuli. Therefore, temporal processing deficit in schizophrenia can lead to hyperintegration of temporally unmatched multisensory stimuli.

  17. A possible approach to optimization of parameters of sound-absorbing structures for multimode waveguides

    NASA Astrophysics Data System (ADS)

    Mironov, M. A.

    2011-11-01

    A method of allowing for the spatial sound field structure in designing the sound-absorbing structures for turbojet aircraft engine ducts is proposed. The acoustic impedance of a duct should be chosen so as to prevent the reflection of the primary sound field, which is generated by the sound source in the absence of the duct, from the duct walls.

  18. Quantifying the influence of flow asymmetries on glottal sound sources in speech

    NASA Astrophysics Data System (ADS)

    Erath, Byron; Plesniak, Michael

    2008-11-01

    Human speech is made possible by the air flow interaction with the vocal folds. During phonation, asymmetries in the glottal flow field may arise from flow phenomena (e.g. the Coanda effect) as well as from pathological vocal fold motion (e.g. unilateral paralysis). In this study, the effects of flow asymmetries on glottal sound sources were investigated. Dynamically-programmable 7.5 times life-size vocal fold models with 2 degrees-of-freedom (linear and rotational) were constructed to provide a first-order approximation of vocal fold motion. Important parameters (Reynolds, Strouhal, and Euler numbers) were scaled to physiological values. Normal and abnormal vocal fold motions were synthesized, and the velocity field and instantaneous transglottal pressure drop were measured. Variability in the glottal jet trajectory necessitated sorting of the data according to the resulting flow configuration. The dipole sound source is related to the transglottal pressure drop via acoustic analogies. Variations in the transglottal pressure drop (and subsequently the dipole sound source) arising from flow asymmetries are discussed.

  19. Auditory event perception: the source-perception loop for posture in human gait.

    PubMed

    Pastore, Richard E; Flint, Jesse D; Gaston, Jeremy R; Solomon, Matthew J

    2008-01-01

    There is a small but growing literature on the perception of natural acoustic events, but few attempts have been made to investigate complex sounds not systematically controlled within a laboratory setting. The present study investigates listeners' ability to make judgments about the posture (upright-stooped) of the walker who generated acoustic stimuli contrasted on each trial. We use a comprehensive three-stage approach to event perception, in which we develop a solid understanding of the source event and its sound properties, as well as the relationships between these two event stages. Developing this understanding helps both to identify the limitations of common statistical procedures and to develop effective new procedures for investigating not only the two information stages above, but also the decision strategies employed by listeners in making source judgments from sound. The result is a comprehensive, ultimately logical, but not necessarily expected picture of both the source-sound-perception loop and the utility of alternative research tools.

  20. Nonlinear theory of shocked sound propagation in a nearly choked duct flow

    NASA Technical Reports Server (NTRS)

    Myers, M. K.; Callegari, A. J.

    1982-01-01

    The development of shocks in the sound field propagating through a nearly choked duct flow is analyzed by extending a quasi-one dimensional theory. The theory is applied to the case in which sound is introduced into the flow by an acoustic source located in the vicinity of a near-sonic throat. Analytical solutions for the field are obtained which illustrate the essential features of the nonlinear interaction between sound and flow. Numerical results are presented covering ranges of variation of source strength, throat Mach number, and frequency. It is found that the development of shocks leads to appreciable attenuation of acoustic power transmitted upstream through the near-sonic flow. It is possible, for example, that the power loss in the fundamental harmonic can be as much as 90% of that introduced at the source.

  1. Noise abatement in a pine plantation

    Treesearch

    R. E. Leonard; L. P. Herrington

    1971-01-01

    Observations on sound propagation were made in two red pine plantations. Measurements were taken of attenuation of prerecorded frequencies at various distances from the sound source. Sound absorption was strongly dependent on frequencies. Peak absorption was at 500 Hz.

  2. Hearing in three dimensions: Sound localization

    NASA Technical Reports Server (NTRS)

    Wightman, Frederic L.; Kistler, Doris J.

    1990-01-01

    The ability to localize a source of sound in space is a fundamental component of the three dimensional character of the sound of audio. For over a century scientists have been trying to understand the physical and psychological processes and physiological mechanisms that subserve sound localization. This research has shown that important information about sound source position is provided by interaural differences in time of arrival, interaural differences in intensity and direction-dependent filtering provided by the pinnae. Progress has been slow, primarily because experiments on localization are technically demanding. Control of stimulus parameters and quantification of the subjective experience are quite difficult problems. Recent advances, such as the ability to simulate a three dimensional sound field over headphones, seem to offer potential for rapid progress. Research using the new techniques has already produced new information. It now seems that interaural time differences are a much more salient and dominant localization cue than previously believed.

  3. Using sounds for making decisions: greater tube-nosed bats prefer antagonistic calls over non-communicative sounds when feeding

    PubMed Central

    Jiang, Tinglei; Long, Zhenyu; Ran, Xin; Zhao, Xue; Xu, Fei; Qiu, Fuyuan; Kanwal, Jagmeet S.

    2016-01-01

    ABSTRACT Bats vocalize extensively within different social contexts. The type and extent of information conveyed via their vocalizations and their perceptual significance, however, remains controversial and difficult to assess. Greater tube-nosed bats, Murina leucogaster, emit calls consisting of long rectangular broadband noise burst (rBNBl) syllables during aggression between males. To experimentally test the behavioral impact of these sounds for feeding, we deployed an approach and place-preference paradigm. Two food trays were placed on opposite sides and within different acoustic microenvironments, created by sound playback, within a specially constructed tent. Specifically, we tested whether the presence of rBNBl sounds at a food source effectively deters the approach of male bats in comparison to echolocation sounds and white noise. In each case, contrary to our expectation, males preferred to feed at a location where rBNBl sounds were present. We propose that the species-specific rBNBl provides contextual information, not present within non-communicative sounds, to facilitate approach towards a food source. PMID:27815241

  4. What the Toadfish Ear Tells the Toadfish Brain About Sound.

    PubMed

    Edds-Walton, Peggy L

    2016-01-01

    Of the three, paired otolithic endorgans in the ear of teleost fishes, the saccule is the one most often demonstrated to have a major role in encoding frequencies of biologically relevant sounds. The toadfish saccule also encodes sound level and sound source direction in the phase-locked activity conveyed via auditory afferents to nuclei of the ipsilateral octaval column in the medulla. Although paired auditory receptors are present in teleost fishes, binaural processes were believed to be unimportant due to the speed of sound in water and the acoustic transparency of the tissues in water. In contrast, there are behavioral and anatomical data that support binaural processing in fishes. Studies in the toadfish combined anatomical tract-tracing and physiological recordings from identified sites along the ascending auditory pathway to document response characteristics at each level. Binaural computations in the medulla and midbrain sharpen the directional information provided by the saccule. Furthermore, physiological studies in the central nervous system indicated that encoding frequency, sound level, temporal pattern, and sound source direction are important components of what the toadfish ear tells the toadfish brain about sound.

  5. Replacing the Orchestra? – The Discernibility of Sample Library and Live Orchestra Sounds

    PubMed Central

    Wolf, Anna; Platz, Friedrich; Mons, Jan

    2016-01-01

    Recently, musical sounds from pre-recorded orchestra sample libraries (OSL) have become indispensable in music production for the stage or popular charts. Surprisingly, it is unknown whether human listeners can identify sounds as stemming from real orchestras or OSLs. Thus, an internet-based experiment was conducted to investigate whether a classic orchestral work, produced with sounds from a state-of-the-art OSL, could be reliably discerned from a live orchestra recording of the piece. It could be shown that the entire sample of listeners (N = 602) on average identified the correct sound source at 72.5%. This rate slightly exceeded Alan Turing's well-known upper threshold of 70% for a convincing, simulated performance. However, while sound experts tended to correctly identify the sound source, participants with lower listening expertise, who resembled the majority of music consumers, only achieved 68.6%. As non-expert listeners in the experiment were virtually unable to tell the real-life and OSL sounds apart, it is assumed that OSLs will become more common in music production for economic reasons. PMID:27382932

  6. The Coast Artillery Journal. Volume 65, Number 4, October 1926

    DTIC Science & Technology

    1926-10-01

    sound. a. Sound location of airplanes by binaural observation in all antiaircraft regiments. b. Sound ranging on report of enemy guns, together with...Direction finding by binaural observation. [Subparagraphs 30 a and 30 c (l).J This applies to continuous sounds such as pro- pellor noises. b. Point...impacts. 32. The so-called binaural sense is our means of sensing the direc- tion of a sound source. When we hear a sound we judge the approxi- mate

  7. Hemispherical breathing mode speaker using a dielectric elastomer actuator.

    PubMed

    Hosoya, Naoki; Baba, Shun; Maeda, Shingo

    2015-10-01

    Although indoor acoustic characteristics should ideally be assessed by measuring the reverberation time using a point sound source, a regular polyhedron loudspeaker, which has multiple loudspeakers on a chassis, is typically used. However, such a configuration is not a point sound source if the size of the loudspeaker is large relative to the target sound field. This study investigates a small lightweight loudspeaker using a dielectric elastomer actuator vibrating in the breathing mode (the pulsating mode such as the expansion and contraction of a balloon). Acoustic testing with regard to repeatability, sound pressure, vibration mode profiles, and acoustic radiation patterns indicate that dielectric elastomer loudspeakers may be feasible.

  8. The role of reverberation-related binaural cues in the externalization of speech.

    PubMed

    Catic, Jasmina; Santurette, Sébastien; Dau, Torsten

    2015-08-01

    The perception of externalization of speech sounds was investigated with respect to the monaural and binaural cues available at the listeners' ears in a reverberant environment. Individualized binaural room impulse responses (BRIRs) were used to simulate externalized sound sources via headphones. The measured BRIRs were subsequently modified such that the proportion of the response containing binaural vs monaural information was varied. Normal-hearing listeners were presented with speech sounds convolved with such modified BRIRs. Monaural reverberation cues were found to be sufficient for the externalization of a lateral sound source. In contrast, for a frontal source, an increased amount of binaural cues from reflections was required in order to obtain well externalized sound images. It was demonstrated that the interaction between the interaural cues of the direct sound and the reverberation strongly affects the perception of externalization. An analysis of the short-term binaural cues showed that the amount of fluctuations of the binaural cues corresponded well to the externalization ratings obtained in the listening tests. The results further suggested that the precedence effect is involved in the auditory processing of the dynamic binaural cues that are utilized for externalization perception.

  9. A stepped-plate bi-frequency source for generating a difference frequency sound with a parametric array.

    PubMed

    Je, Yub; Lee, Haksue; Park, Jongkyu; Moon, Wonkyu

    2010-06-01

    An ultrasonic radiator is developed to generate a difference frequency sound from two frequencies of ultrasound in air with a parametric array. A design method is proposed for an ultrasonic radiator capable of generating highly directive, high-amplitude ultrasonic sound beams at two different frequencies in air based on a modification of the stepped-plate ultrasonic radiator. The stepped-plate ultrasonic radiator was introduced by Gallego-Juarez et al. [Ultrasonics 16, 267-271 (1978)] in their previous study and can effectively generate highly directive, large-amplitude ultrasonic sounds in air, but only at a single frequency. Because parametric array sources must be able to generate sounds at more than one frequency, a design modification is crucial to the application of a stepped-plate ultrasonic radiator as a parametric array source in air. The aforementioned method was employed to design a parametric radiator for use in air. A prototype of this design was constructed and tested to determine whether it could successfully generate a difference frequency sound with a parametric array. The results confirmed that the proposed single small-area transducer was suitable as a parametric radiator in air.

  10. Chronic detachable headphones for acoustic stimulation in freely moving animals.

    PubMed

    Nodal, Fernando R; Keating, Peter; King, Andrew J

    2010-05-30

    A growing number of studies of auditory processing are being carried out in awake, behaving animals, creating a need for precisely controlled sound delivery without restricting head movements. We have designed a system for closed-field stimulus presentation in freely moving ferrets, which comprises lightweight, adjustable headphones that can be consistently positioned over the ears via a small, skull-mounted implant. The invasiveness of the implant was minimized by simplifying its construction and using dental adhesive only for attaching it to the skull, thereby reducing the surgery required and avoiding the use of screws or other anchoring devices. Attaching the headphones to a chronic implant also reduced the amount of contact they had with the head and ears, increasing the willingness of the animals to wear them. We validated sound stimulation via the headphones in ferrets trained previously in a free-field task to localize stimuli presented from one of two loudspeakers. Noise bursts were delivered binaurally over the headphones and interaural level differences (ILDs) were introduced to allow the sound to be lateralized. Animals rapidly transferred from the free-field task to indicate the perceived location of the stimulus presented over headphones. They showed near perfect lateralization with a 5 dB ILD, matching the scores achieved in the free-field task. As expected, the ferrets' performance declined when the ILD was reduced in value. This closed-field system can easily be adapted for use in other species, and provides a reliable means of presenting closed-field stimuli whilst monitoring behavioral responses in freely moving animals. (c) 2010 Elsevier B.V. All rights reserved.

  11. Bearings Only Air-to-Air Ranging

    DTIC Science & Technology

    1988-07-25

    directly in fiut of the observer whem first detected, more time will be needed for a good estimate. A sound uinp them is for the observer, having...altitude angle to provide an estimate of the z component. Moving targets commonly require some 60 seconds for good estimates of target location and...fixed target case, where a good strategy for the observer can be determined a priori, highly effective maneuvers for the observer in the case of a moving

  12. On the hazard of quiet vehicles to pedestrians and drivers.

    PubMed

    Wogalter, Michael S; Lim, Raymond W; Nyeste, Patrick G

    2014-09-01

    The need to produce more efficient and less polluting vehicles has encouraged mass production of alternative energy vehicles, such as hybrid and electric cars. Many of these vehicles are capable of very quiet operation. While reducing noise pollution is desirable, quieter vehicles could negatively affect pedestrian safety because of reduced sound cues compared to louder internal combustion engines. Three studies were performed to investigate people's concern about this issue. In Study 1, a questionnaire completed by 378 people showed substantial positive interest in quiet hybrid and electric cars. However, they also indicated concern about the reduced auditory cues of quiet vehicles. In Study 2, 316 participants rated 14 sounds that could be potentially added to quiet alternative-energy vehicles. The data showed that participants did not want annoying sounds, but preferred adding "engine" and "hum" sounds relative to other types of sounds. In Study 3, 24 persons heard and rated 18 actual sounds within 6 categories that were added to a video of a hybrid vehicle driving by. The sounds most preferred were "engine" followed by "white noise" and "hum". Implications for adding sounds to facilitate pedestrians' detection of moving vehicles and for aiding drivers' awareness of speed are discussed. Copyright © 2013 Elsevier Ltd and The Ergonomics Society. All rights reserved.

  13. An open access database for the evaluation of heart sound algorithms.

    PubMed

    Liu, Chengyu; Springer, David; Li, Qiao; Moody, Benjamin; Juan, Ricardo Abad; Chorro, Francisco J; Castells, Francisco; Roig, José Millet; Silva, Ikaro; Johnson, Alistair E W; Syed, Zeeshan; Schmidt, Samuel E; Papadaniil, Chrysa D; Hadjileontiadis, Leontios; Naseri, Hosein; Moukadem, Ali; Dieterlen, Alain; Brandt, Christian; Tang, Hong; Samieinasab, Maryam; Samieinasab, Mohammad Reza; Sameni, Reza; Mark, Roger G; Clifford, Gari D

    2016-12-01

    In the past few decades, analysis of heart sound signals (i.e. the phonocardiogram or PCG), especially for automated heart sound segmentation and classification, has been widely studied and has been reported to have the potential value to detect pathology accurately in clinical applications. However, comparative analyses of algorithms in the literature have been hindered by the lack of high-quality, rigorously validated, and standardized open databases of heart sound recordings. This paper describes a public heart sound database, assembled for an international competition, the PhysioNet/Computing in Cardiology (CinC) Challenge 2016. The archive comprises nine different heart sound databases sourced from multiple research groups around the world. It includes 2435 heart sound recordings in total collected from 1297 healthy subjects and patients with a variety of conditions, including heart valve disease and coronary artery disease. The recordings were collected from a variety of clinical or nonclinical (such as in-home visits) environments and equipment. The length of recording varied from several seconds to several minutes. This article reports detailed information about the subjects/patients including demographics (number, age, gender), recordings (number, location, state and time length), associated synchronously recorded signals, sampling frequency and sensor type used. We also provide a brief summary of the commonly used heart sound segmentation and classification methods, including open source code provided concurrently for the Challenge. A description of the PhysioNet/CinC Challenge 2016, including the main aims, the training and test sets, the hand corrected annotations for different heart sound states, the scoring mechanism, and associated open source code are provided. In addition, several potential benefits from the public heart sound database are discussed.

  14. Neuromagnetic recordings reveal the temporal dynamics of auditory spatial processing in the human cortex.

    PubMed

    Tiitinen, Hannu; Salminen, Nelli H; Palomäki, Kalle J; Mäkinen, Ville T; Alku, Paavo; May, Patrick J C

    2006-03-20

    In an attempt to delineate the assumed 'what' and 'where' processing streams, we studied the processing of spatial sound in the human cortex by using magnetoencephalography in the passive and active recording conditions and two kinds of spatial stimuli: individually constructed, highly realistic spatial (3D) stimuli and stimuli containing interaural time difference (ITD) cues only. The auditory P1m, N1m, and P2m responses of the event-related field were found to be sensitive to the direction of sound source in the azimuthal plane. In general, the right-hemispheric responses to spatial sounds were more prominent than the left-hemispheric ones. The right-hemispheric P1m and N1m responses peaked earlier for sound sources in the contralateral than for sources in the ipsilateral hemifield and the peak amplitudes of all responses reached their maxima for contralateral sound sources. The amplitude of the right-hemispheric P2m response reflected the degree of spatiality of sound, being twice as large for the 3D than ITD stimuli. The results indicate that the right hemisphere is specialized in the processing of spatial cues in the passive recording condition. Minimum current estimate (MCE) localization revealed that temporal areas were activated both in the active and passive condition. This initial activation, taking place at around 100 ms, was followed by parietal and frontal activity at 180 and 200 ms, respectively. The latter activations, however, were specific to attentional engagement and motor responding. This suggests that parietal activation reflects active responding to a spatial sound rather than auditory spatial processing as such.

  15. Sound Source Localization through 8 MEMS Microphones Array Using a Sand-Scorpion-Inspired Spiking Neural Network.

    PubMed

    Beck, Christoph; Garreau, Guillaume; Georgiou, Julius

    2016-01-01

    Sand-scorpions and many other arachnids perceive their environment by using their feet to sense ground waves. They are able to determine amplitudes the size of an atom and locate the acoustic stimuli with an accuracy of within 13° based on their neuronal anatomy. We present here a prototype sound source localization system, inspired from this impressive performance. The system presented utilizes custom-built hardware with eight MEMS microphones, one for each foot, to acquire the acoustic scene, and a spiking neural model to localize the sound source. The current implementation shows smaller localization error than those observed in nature.

  16. The influence of crowd density on the sound environment of commercial pedestrian streets.

    PubMed

    Meng, Qi; Kang, Jian

    2015-04-01

    Commercial pedestrian streets are very common in China and Europe, with many situated in historic or cultural centres. The environments of these streets are important, including their sound environments. The objective of this study is to explore the relationships between the crowd density and the sound environments of commercial pedestrian streets. On-site measurements were performed at the case study site in Harbin, China, and a questionnaire was administered. The sound pressure measurements showed that the crowd density has an insignificant effect on sound pressure below 0.05 persons/m2, whereas when the crowd density is greater than 0.05 persons/m2, the sound pressure increases with crowd density. The sound sources were analysed, showing that several typical sound sources, such as traffic noise, can be masked by the sounds resulting from dense crowds. The acoustic analysis showed that crowd densities outside the range of 0.10 to 0.25 persons/m2 exhibited lower acoustic comfort evaluation scores. In terms of audiovisual characteristics, the subjective loudness increases with greater crowd density, while the acoustic comfort decreases. The results for an indoor underground shopping street are also presented for comparison. Copyright © 2014 Elsevier B.V. All rights reserved.

  17. Soundscapes and the sense of hearing of fishes.

    PubMed

    Fay, Richard

    2009-03-01

    Underwater soundscapes have probably played an important role in the adaptation of ears and auditory systems of fishes throughout evolutionary time, and for all species. These sounds probably contain important information about the environment and about most objects and events that confront the receiving fish so that appropriate behavior is possible. For example, the sounds from reefs appear to be used by at least some fishes for their orientation and migration. These sorts of environmental sounds should be considered much like "acoustic daylight," that continuously bathes all environments and contain information that all organisms can potentially use to form a sort of image of the environment. At present, however, we are generally ignorant of the nature of ambient sound fields impinging on fishes, and the adaptive value of processing these fields to resolve the multiple sources of sound. Our field has focused almost exclusively on the adaptive value of processing species-specific communication sounds, and has not considered the informational value of ambient "noise." Since all fishes can detect and process acoustic particle motion, including the directional characteristics of this motion, underwater sound fields are potentially more complex and information-rich than terrestrial acoustic environments. The capacities of one fish species (goldfish) to receive and make use of such sound source information have been demonstrated (sound source segregation and auditory scene analysis), and it is suggested that all vertebrate species have this capacity. A call is made to better understand underwater soundscapes, and the associated behaviors they determine in fishes. © 2009 ISZS, Blackwell Publishing and IOZ/CAS.

  18. Possibilities of psychoacoustics to determine sound quality

    NASA Astrophysics Data System (ADS)

    Genuit, Klaus

    For some years, acoustic engineers have increasingly become aware of the importance of analyzing and minimizing noise problems not only with regard to the A-weighted sound pressure level, but to design sound quality. It is relatively easy to determine the A-weighted SPL according to international standards. However, the objective evaluation to describe subjectively perceived sound quality - taking into account psychoacoustic parameters such as loudness, roughness, fluctuation strength, sharpness and so forth - is more difficult. On the one hand, the psychoacoustic measurement procedures which are known so far have yet not been standardized. On the other hand, they have only been tested in laboratories by means of listening tests in the free-field and one sound source and simple signals. Therefore, the results achieved cannot be transferred to complex sound situations with several spatially distributed sound sources without difficulty. Due to the directional hearing and selectivity of human hearing, individual sound events can be selected among many. Already in the late seventies a new binaural Artificial Head Measurement System was developed which met the requirements of the automobile industry in terms of measurement technology. The first industrial application of the Artificial Head Measurement System was in 1981. Since that time the system was further developed, particularly by the cooperation between HEAD acoustics and Mercedes-Benz. In addition to a calibratable Artificial Head Measurement System which is compatible with standard measurement technologies and has transfer characteristics comparable to human hearing, a Binaural Analysis System is now also available. This system permits the analysis of binaural signals regarding physical and psychoacoustic procedures. Moreover, the signals to be analyzed can be simultaneously monitored through headphones and manipulated in the time and frequency domain so that those signal components being responsible for noise annoyance can be found. Especially in complex sound situations with several spatially distributed sound sources, standard, one-channel measurements methods cannot adequately determine sound quality, the acoustic comfort, or annoyance of sound events.

  19. Aerial hawking and landing: approach behaviour in Natterer's bats, Myotis nattereri (Kuhl 1818).

    PubMed

    Melcón, Mariana L; Denzinger, Annette; Schnitzler, Hans-Ulrich

    2007-12-01

    We compared the flight and echolocation behaviour of a vespertilionid bat (Myotis nattereri) approaching a large stationary or a small moving target. Bats were trained to either land on a landing grid or to catch a moving tethered mealworm. When closing in on these two targets, the bats emitted groups of sounds with increasing number of signals and decreasing pulse interval and duration. When pursuing the mealworm, the approach phase always ended with a terminal group consisting of buzz I and buzz II. When landing, the bats emitted either a terminal group consisting of buzz I alone, with one or two extra pulses, or a group consisting of buzz I and buzz II. In all situations, buzz I ended on average between 47-63 ms prior to contact with the target of interest, which is approximately the reaction time of bats. Therefore, the information collected in buzz II does not guide the bats to the target. The relevant part of the approach phase to reach the target ends with buzz I. The basic sound pattern of this part is rather similar and independent of whether the bats approach the large stationary or the small moving target.

  20. Using ILD or ITD Cues for Sound Source Localization and Speech Understanding in a Complex Listening Environment by Listeners With Bilateral and With Hearing-Preservation Cochlear Implants.

    PubMed

    Loiselle, Louise H; Dorman, Michael F; Yost, William A; Cook, Sarah J; Gifford, Rene H

    2016-08-01

    To assess the role of interaural time differences and interaural level differences in (a) sound-source localization, and (b) speech understanding in a cocktail party listening environment for listeners with bilateral cochlear implants (CIs) and for listeners with hearing-preservation CIs. Eleven bilateral listeners with MED-EL (Durham, NC) CIs and 8 listeners with hearing-preservation CIs with symmetrical low frequency, acoustic hearing using the MED-EL or Cochlear device were evaluated using 2 tests designed to task binaural hearing, localization, and a simulated cocktail party. Access to interaural cues for localization was constrained by the use of low-pass, high-pass, and wideband noise stimuli. Sound-source localization accuracy for listeners with bilateral CIs in response to the high-pass noise stimulus and sound-source localization accuracy for the listeners with hearing-preservation CIs in response to the low-pass noise stimulus did not differ significantly. Speech understanding in a cocktail party listening environment improved for all listeners when interaural cues, either interaural time difference or interaural level difference, were available. The findings of the current study indicate that similar degrees of benefit to sound-source localization and speech understanding in complex listening environments are possible with 2 very different rehabilitation strategies: the provision of bilateral CIs and the preservation of hearing.

  1. Spherical harmonic analysis of the sound radiation from omnidirectional loudspeaker arrays

    NASA Astrophysics Data System (ADS)

    Pasqual, A. M.

    2014-09-01

    Omnidirectional sound sources are widely used in room acoustics. These devices are made up of loudspeakers mounted on a spherical or polyhedral cabinet, where the dodecahedral shape prevails. Although such electroacoustic sources have been made readily available to acousticians by many manufacturers, an in-depth investigation of their vibroacoustic behavior has not been provided yet. In order to fulfill this lack, this paper presents a theoretical study of the sound radiation from omnidirectional loudspeaker arrays, which is carried out by using a mathematical model based on the spherical harmonic analysis. Eight different loudspeaker arrangements on the sphere are considered: the well-known five Platonic solid layouts and three extremal system layouts. The latter possess useful properties for spherical loudspeaker arrays used as directivity controlled sound sources, so that these layouts are included here in order to investigate whether or not they could be of interest as omnidirectional sources as well. It is shown through a comparative analysis that the dodecahedral array leads to the lowest error in producing an omnidirectional sound field and to the highest acoustic power, which corroborates the prevalence of such a layout. In addition, if a source with less than 12 loudspeakers is required, it is shown that tetrahedra or hexahedra can be used alternatively, whereas the extremal system layouts are not interesting choices for omnidirectional loudspeaker arrays.

  2. The use of an active controlled enclosure to attenuate sound radiation from a heavy radiator

    NASA Astrophysics Data System (ADS)

    Sun, Yao; Yang, Tiejun; Zhu, Minggang; Pan, Jie

    2017-03-01

    Active structural acoustical control usually experiences difficulty in the control of heavy sources or sources where direct applications of control forces are not practical. To overcome this difficulty, an active controlled enclosure, which forms a cavity with both flexible and open boundary, is employed. This configuration permits indirect implementation of active control in which the control inputs can be applied to subsidiary structures other than the sources. To determine the control effectiveness of the configuration, the vibro-acoustic behavior of the system, which consists of a top plate with an open, a sound cavity and a source panel, is investigated in this paper. A complete mathematical model of the system is formulated involving modified Fourier series formulations and the governing equations are solved using the Rayleigh-Ritz method. The coupling mechanisms of a partly opened cavity and a plate are analysed in terms of modal responses and directivity patterns. Furthermore, to attenuate sound power radiated from both the top panel and the open, two strategies are studied: minimizing the total radiated power and the cancellation of volume velocity. Moreover, three control configurations are compared, using a point force on the control panel (structural control), using a sound source in the cavity (acoustical control) and applying hybrid structural-acoustical control. In addition, the effects of boundary condition of the control panel on the sound radiation and control performance are discussed.

  3. Material sound source localization through headphones

    NASA Astrophysics Data System (ADS)

    Dunai, Larisa; Peris-Fajarnes, Guillermo; Lengua, Ismael Lengua; Montaña, Ignacio Tortajada

    2012-09-01

    In the present paper a study of sound localization is carried out, considering two different sounds emitted from different hit materials (wood and bongo) as well as a Delta sound. The motivation of this research is to study how humans localize sounds coming from different materials, with the purpose of a future implementation of the acoustic sounds with better localization features in navigation aid systems or training audio-games suited for blind people. Wood and bongo sounds are recorded after hitting two objects made of these materials. Afterwards, they are analysed and processed. On the other hand, the Delta sound (click) is generated by using the Adobe Audition software, considering a frequency of 44.1 kHz. All sounds are analysed and convolved with previously measured non-individual Head-Related Transfer Functions both for an anechoic environment and for an environment with reverberation. The First Choice method is used in this experiment. Subjects are asked to localize the source position of the sound listened through the headphones, by using a graphic user interface. The analyses of the recorded data reveal that no significant differences are obtained either when considering the nature of the sounds (wood, bongo, Delta) or their environmental context (with or without reverberation). The localization accuracies for the anechoic sounds are: wood 90.19%, bongo 92.96% and Delta sound 89.59%, whereas for the sounds with reverberation the results are: wood 90.59%, bongo 92.63% and Delta sound 90.91%. According to these data, we can conclude that even when considering the reverberation effect, the localization accuracy does not significantly increase.

  4. Aeroacoustic analysis of the human phonation process based on a hybrid acoustic PIV approach

    NASA Astrophysics Data System (ADS)

    Lodermeyer, Alexander; Tautz, Matthias; Becker, Stefan; Döllinger, Michael; Birk, Veronika; Kniesburges, Stefan

    2018-01-01

    The detailed analysis of sound generation in human phonation is severely limited as the accessibility to the laryngeal flow region is highly restricted. Consequently, the physical basis of the underlying fluid-structure-acoustic interaction that describes the primary mechanism of sound production is not yet fully understood. Therefore, we propose the implementation of a hybrid acoustic PIV procedure to evaluate aeroacoustic sound generation during voice production within a synthetic larynx model. Focusing on the flow field downstream of synthetic, aerodynamically driven vocal folds, we calculated acoustic source terms based on the velocity fields obtained by time-resolved high-speed PIV applied to the mid-coronal plane. The radiation of these sources into the acoustic far field was numerically simulated and the resulting acoustic pressure was finally compared with experimental microphone measurements. We identified the tonal sound to be generated downstream in a small region close to the vocal folds. The simulation of the sound propagation underestimated the tonal components, whereas the broadband sound was well reproduced. Our results demonstrate the feasibility to locate aeroacoustic sound sources inside a synthetic larynx using a hybrid acoustic PIV approach. Although the technique employs a 2D-limited flow field, it accurately reproduces the basic characteristics of the aeroacoustic field in our larynx model. In future studies, not only the aeroacoustic mechanisms of normal phonation will be assessable, but also the sound generation of voice disorders can be investigated more profoundly.

  5. Selective attention to sound location or pitch studied with event-related brain potentials and magnetic fields.

    PubMed

    Degerman, Alexander; Rinne, Teemu; Särkkä, Anna-Kaisa; Salmi, Juha; Alho, Kimmo

    2008-06-01

    Event-related brain potentials (ERPs) and magnetic fields (ERFs) were used to compare brain activity associated with selective attention to sound location or pitch in humans. Sixteen healthy adults participated in the ERP experiment, and 11 adults in the ERF experiment. In different conditions, the participants focused their attention on a designated sound location or pitch, or pictures presented on a screen, in order to detect target sounds or pictures among the attended stimuli. In the Attend Location condition, the location of sounds varied randomly (left or right), while their pitch (high or low) was kept constant. In the Attend Pitch condition, sounds of varying pitch (high or low) were presented at a constant location (left or right). Consistent with previous ERP results, selective attention to either sound feature produced a negative difference (Nd) between ERPs to attended and unattended sounds. In addition, ERPs showed a more posterior scalp distribution for the location-related Nd than for the pitch-related Nd, suggesting partially different generators for these Nds. The ERF source analyses found no source distribution differences between the pitch-related Ndm (the magnetic counterpart of the Nd) and location-related Ndm in the superior temporal cortex (STC), where the main sources of the Ndm effects are thought to be located. Thus, the ERP scalp distribution differences between the location-related and pitch-related Nd effects may have been caused by activity of areas outside the STC, perhaps in the inferior parietal regions.

  6. Magnetotellurics with long distance remote reference to reject DC railway noise

    NASA Astrophysics Data System (ADS)

    Hanstein, T.; Jiang, J.; Strack, K.; Ritter, O.

    2014-12-01

    Some parts of railway network in Europe is electrified by DC current. The return current in the ground is varying in space, time and power when the train is moving. Since the train traffic is active 24 hours, there is no quite time. The train signal is dominating for periods longer than 1 s and is a near field source. The transfer function of the magnetotelluric sounding (MT) is influenced by this near field source, the phase is going to zero and amplitude increase with slope 1 for longer periods. Since this dominating noise is present all day robust magnetotelluric processing technique to identify and remove outliers are not applicable and sufficient. The remote reference technique has successfully been applied for magnetotelluric soundings Combining an disturbed local MT data set with the data of the remote station, which is recording simultaneously the horizontal magnetic fields, can improve the data quality. Finding a good remote station during field survey is difficult and expensive. There is a permanent MT remote reference station in Germany. The set up and maintance is done by the GFZ - Helmholtz Centre Potsdam - GFZ German Research Centre for Geosciences. The location is near Wittstock and has good signal-to-noise-ratio with low cutural noise, the ground is almost lD and recording since May 2010. The electric and magnetic field is continously recorded with 250 Hz sampling and induction coils. The magnetic field is also recorded with fluxgate magnetometers and 5 Hz sampling. The distance to the local MT site is about 600 km.

  7. Sound Explorations from the Ages of 10 to 37 Months: The Ontogenesis of Musical Conducts

    ERIC Educational Resources Information Center

    Delalande, Francois; Cornara, Silvia

    2010-01-01

    One of the forms of first musical conduct is the exploration of sound sources. When young children produce sounds with any object, these sounds may surprise them and so they make the sounds again--not exactly the same, but introducing some variation. A process of repetition with slight changes is set in motion which can be analysed, as did Piaget,…

  8. Monitoring the Ocean Using High Frequency Ambient Sound

    DTIC Science & Technology

    2008-10-01

    even identify specific groups within the resident killer whale type ( Puget Sound Southern Resident pods J, K and L) because these groups have...particular, the different populations of killer whales in the NE Pacific Ocean. This has been accomplished by detecting transient sounds with short...high sea state (the sound of spray), general shipping - close and distant, clanking and whale calls and clicking. These sound sources form the basis

  9. Prospective cohort study on noise levels in a pediatric cardiac intensive care unit.

    PubMed

    Garcia Guerra, Gonzalo; Joffe, Ari R; Sheppard, Cathy; Pugh, Jodie; Moez, Elham Khodayari; Dinu, Irina A; Jou, Hsing; Hartling, Lisa; Vohra, Sunita

    2018-04-01

    To describe noise levels in a pediatric cardiac intensive care unit, and to determine the relationship between sound levels and patient sedation requirements. Prospective observational study at a pediatric cardiac intensive care unit (PCICU). Sound levels were measured continuously in slow A weighted decibels dB(A) with a sound level meter SoundEarPro® during a 4-week period. Sedation requirement was assessed using the number of intermittent (PRNs) doses given per hour. Analysis was conducted with autoregressive moving average models and the Granger test for causality. 39 children were included in the study. The average (SD) sound level in the open area was 59.4 (2.5) dB(A) with a statistically significant but clinically unimportant difference between day/night hours (60.1 vs. 58.6; p-value < 0.001). There was no significant difference between sound levels in the open area/single room (59.4 vs. 60.8, p-value = 0.108). Peak noise levels were > 90 dB. There was a significant association between average (p-value = 0.030) and peak sound levels (p-value = 0.006), and number of sedation PRNs. Sound levels were above the recommended values with no differences between day/night or open area/single room. High sound levels were significantly associated with sedation requirements. Copyright © 2017 Elsevier Inc. All rights reserved.

  10. Slip-stick excitation and travelling waves excite silo honking

    NASA Astrophysics Data System (ADS)

    Warburton, Katarzyna; Porte, Elze; Vriend, Nathalie

    2017-06-01

    Silo honking is the harmonic sound generated by the discharge of a silo filled with a granular material. In industrial storage silos, the acoustic emission during discharge of PET-particles forms a nuisance for the environment and may ultimately result in structural failure. This work investigates the phenomenon experimentally using a laboratory-scale silo, and successfully correlates the frequency of the emitted sound with the periodicity of the mechanical motion of the grains. The key driver is the slip-stick interaction between the wall and the particles, characterized as a wave moving upwards through the silo. A quantitative correlation is established for the first time between the frequency of the sound, measured with an electret microphone, and the slip-frequency, measured with a high-speed camera. In the lower regions of the tube, both the slip-stick motion and the honking sound disappear.

  11. Acoustic topological insulator and robust one-way sound transport

    NASA Astrophysics Data System (ADS)

    He, Cheng; Ni, Xu; Ge, Hao; Sun, Xiao-Chen; Chen, Yan-Bin; Lu, Ming-Hui; Liu, Xiao-Ping; Chen, Yan-Feng

    2016-12-01

    Topological design of materials enables topological symmetries and facilitates unique backscattering-immune wave transport. In airborne acoustics, however, the intrinsic longitudinal nature of sound polarization makes the use of the conventional spin-orbital interaction mechanism impossible for achieving band inversion. The topological gauge flux is then typically introduced with a moving background in theoretical models. Its practical implementation is a serious challenge, though, due to inherent dynamic instabilities and noise. Here we realize the inversion of acoustic energy bands at a double Dirac cone and provide an experimental demonstration of an acoustic topological insulator. By manipulating the hopping interaction of neighbouring ’atoms’ in this new topological material, we successfully demonstrate the acoustic quantum spin Hall effect, characterized by robust pseudospin-dependent one-way edge sound transport. Our results are promising for the exploration of new routes for experimentally studying topological phenomena and related applications, for example, sound-noise reduction.

  12. Meteorological effects on long-range outdoor sound propagation

    NASA Technical Reports Server (NTRS)

    Klug, Helmut

    1990-01-01

    Measurements of sound propagation over distances up to 1000 m were carried out with an impulse sound source offering reproducible, short time signals. Temperature and wind speed at several heights were monitored simultaneously; the meteorological data are used to determine the sound speed gradients according to the Monin-Obukhov similarity theory. The sound speed profile is compared to a corresponding prediction, gained through the measured travel time difference between direct and ground reflected pulse (which depends on the sound speed gradient). Positive sound speed gradients cause bending of the sound rays towards the ground yielding enhanced sound pressure levels. The measured meteorological effects on sound propagation are discussed and illustrated by ray tracing methods.

  13. Tongue-driven sonar beam steering by a lingual-echolocating fruit bat

    PubMed Central

    Falk, Benjamin; Chiu, Chen; Krishnan, Anand; Arbour, Jessica H.; Moss, Cynthia F.

    2017-01-01

    Animals enhance sensory acquisition from a specific direction by movements of head, ears, or eyes. As active sensing animals, echolocating bats also aim their directional sonar beam to selectively “illuminate” a confined volume of space, facilitating efficient information processing by reducing echo interference and clutter. Such sonar beam control is generally achieved by head movements or shape changes of the sound-emitting mouth or nose. However, lingual-echolocating Egyptian fruit bats, Rousettus aegyptiacus, which produce sound by clicking their tongue, can dramatically change beam direction at very short temporal intervals without visible morphological changes. The mechanism supporting this capability has remained a mystery. Here, we measured signals from free-flying Egyptian fruit bats and discovered a systematic angular sweep of beam focus across increasing frequency. This unusual signal structure has not been observed in other animals and cannot be explained by the conventional and widely-used “piston model” that describes the emission pattern of other bat species. Through modeling, we show that the observed beam features can be captured by an array of tongue-driven sound sources located along the side of the mouth, and that the sonar beam direction can be steered parsimoniously by inducing changes to the pattern of phase differences through moving tongue location. The effects are broadly similar to those found in a phased array—an engineering design widely found in human-made sonar systems that enables beam direction changes without changes in the physical transducer assembly. Our study reveals an intriguing parallel between biology and human engineering in solving problems in fundamentally similar ways. PMID:29244805

  14. Tongue-driven sonar beam steering by a lingual-echolocating fruit bat.

    PubMed

    Lee, Wu-Jung; Falk, Benjamin; Chiu, Chen; Krishnan, Anand; Arbour, Jessica H; Moss, Cynthia F

    2017-12-01

    Animals enhance sensory acquisition from a specific direction by movements of head, ears, or eyes. As active sensing animals, echolocating bats also aim their directional sonar beam to selectively "illuminate" a confined volume of space, facilitating efficient information processing by reducing echo interference and clutter. Such sonar beam control is generally achieved by head movements or shape changes of the sound-emitting mouth or nose. However, lingual-echolocating Egyptian fruit bats, Rousettus aegyptiacus, which produce sound by clicking their tongue, can dramatically change beam direction at very short temporal intervals without visible morphological changes. The mechanism supporting this capability has remained a mystery. Here, we measured signals from free-flying Egyptian fruit bats and discovered a systematic angular sweep of beam focus across increasing frequency. This unusual signal structure has not been observed in other animals and cannot be explained by the conventional and widely-used "piston model" that describes the emission pattern of other bat species. Through modeling, we show that the observed beam features can be captured by an array of tongue-driven sound sources located along the side of the mouth, and that the sonar beam direction can be steered parsimoniously by inducing changes to the pattern of phase differences through moving tongue location. The effects are broadly similar to those found in a phased array-an engineering design widely found in human-made sonar systems that enables beam direction changes without changes in the physical transducer assembly. Our study reveals an intriguing parallel between biology and human engineering in solving problems in fundamentally similar ways.

  15. The Problems with "Noise Numbers" for Wind Farm Noise Assessment

    ERIC Educational Resources Information Center

    Thorne, Bob

    2011-01-01

    Human perception responds primarily to sound character rather than sound level. Wind farms are unique sound sources and exhibit special audible and inaudible characteristics that can be described as modulating sound or as a tonal complex. Wind farm compliance measures based on a specified noise number alone will fail to address problems with noise…

  16. Popcorn: critical temperature, jump and sound.

    PubMed

    Virot, Emmanuel; Ponomarenko, Alexandre

    2015-03-06

    Popcorn bursts open, jumps and emits a 'pop' sound in some hundredths of a second. The physical origin of these three observations remains unclear in the literature. We show that the critical temperature 180°C at which almost all of popcorn pops is consistent with an elementary pressure vessel scenario. We observe that popcorn jumps with a 'leg' of starch which is compressed on the ground. As a result, popcorn is midway between two categories of moving systems: explosive plants using fracture mechanisms and jumping animals using muscles. By synchronizing video recordings with acoustic recordings, we propose that the familiar 'pop' sound of the popcorn is caused by the release of water vapour. © 2015 The Author(s) Published by the Royal Society. All rights reserved.

  17. Acoustic Event Detection and Classification

    NASA Astrophysics Data System (ADS)

    Temko, Andrey; Nadeu, Climent; Macho, Dušan; Malkin, Robert; Zieger, Christian; Omologo, Maurizio

    The human activity that takes place in meeting rooms or classrooms is reflected in a rich variety of acoustic events (AE), produced either by the human body or by objects handled by humans, so the determination of both the identity of sounds and their position in time may help to detect and describe that human activity. Indeed, speech is usually the most informative sound, but other kinds of AEs may also carry useful information, for example, clapping or laughing inside a speech, a strong yawn in the middle of a lecture, a chair moving or a door slam when the meeting has just started. Additionally, detection and classification of sounds other than speech may be useful to enhance the robustness of speech technologies like automatic speech recognition.

  18. Synthesis of information on the effects of noise and disturbance on major haulout concentrations of Bering Sea pinnipeds. Final report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, S.R.; Burns, J.J.; Malme, C.I.

    1989-02-17

    The study investigated the use of terrestrial haulout sites in the eastern Bering Sea by four species of pennipeds, northern fur seal, northern sea lion, harbor seal and pacific walrus. Historical information on the use of each site was summarized. Available information on the effects of airborne and waterborne noise, and human disturbance (from stationary and moving sources) was reviewed. The authors also conducted a detailed analysis of the acoustic environment of eight haulout sites that were representative of others used by each of the four species studied. The analyses included investigations of (1) characteristics airborne and underwater ambient noise,more » (2) characteristics of industrial noise sources, including aircraft, small boats, fishing trawlers and commercial cargo traffic, and (3) sound transmission loss in air, water, and through the air-water surface. As a means to evaluate the potential vulnerability of each haulout site to noise and disturbance, a quantitative rating system (IPSI) whereby an index of sensitivity was assigned to each site.« less

  19. 49 CFR 210.3 - Applicability.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... emitted by moving rail cars and locomotives (including the sound produced by refrigeration and air conditioning units that are an integral element of such equipment), active retarders, switcher locomotives, car... apply to— (1) Steam locomotives; (2) Street, suburban, or interurban electric railways unless operated...

  20. 49 CFR 210.3 - Applicability.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... emitted by moving rail cars and locomotives (including the sound produced by refrigeration and air conditioning units that are an integral element of such equipment), active retarders, switcher locomotives, car... apply to— (1) Steam locomotives; (2) Street, suburban, or interurban electric railways unless operated...

  1. 49 CFR 210.3 - Applicability.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... emitted by moving rail cars and locomotives (including the sound produced by refrigeration and air conditioning units that are an integral element of such equipment), active retarders, switcher locomotives, car... apply to— (1) Steam locomotives; (2) Street, suburban, or interurban electric railways unless operated...

  2. 49 CFR 210.3 - Applicability.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... emitted by moving rail cars and locomotives (including the sound produced by refrigeration and air conditioning units that are an integral element of such equipment), active retarders, switcher locomotives, car... apply to— (1) Steam locomotives; (2) Street, suburban, or interurban electric railways unless operated...

  3. 49 CFR 210.3 - Applicability.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... emitted by moving rail cars and locomotives (including the sound produced by refrigeration and air conditioning units that are an integral element of such equipment), active retarders, switcher locomotives, car... apply to— (1) Steam locomotives; (2) Street, suburban, or interurban electric railways unless operated...

  4. Spatial sound field synthesis and upmixing based on the equivalent source method.

    PubMed

    Bai, Mingsian R; Hsu, Hoshen; Wen, Jheng-Ciang

    2014-01-01

    Given scarce number of recorded signals, spatial sound field synthesis with an extended sweet spot is a challenging problem in acoustic array signal processing. To address the problem, a synthesis and upmixing approach inspired by the equivalent source method (ESM) is proposed. The synthesis procedure is based on the pressure signals recorded by a microphone array and requires no source model. The array geometry can also be arbitrary. Four upmixing strategies are adopted to enhance the resolution of the reproduced sound field when there are more channels of loudspeakers than the microphones. Multi-channel inverse filtering with regularization is exploited to deal with the ill-posedness in the reconstruction process. The distance between the microphone and loudspeaker arrays is optimized to achieve the best synthesis quality. To validate the proposed system, numerical simulations and subjective listening experiments are performed. The results demonstrated that all upmixing methods improved the quality of reproduced target sound field over the original reproduction. In particular, the underdetermined ESM interpolation method yielded the best spatial sound field synthesis in terms of the reproduction error, timbral quality, and spatial quality.

  5. Einstein’s Clocks

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lincoln, Don

    One of the most non-intuitive physics theories ever devised is Einstein’s Theory of Special Relativity, which claim such crazy-sounding things as two people disagreeing on such familiar concepts as length and time. In this video, Fermilab’s Dr. Don Lincoln shows that every single day particle physicists prove that moving clocks tick more slowly than stationary ones. He uses an easy to understand example of particles that move for far longer distances than you would expect from combining their velocity and stationary lifetime.

  6. Development and Testing of a High Level Axial Array Duct Sound Source for the NASA Flow Impedance Test Facility

    NASA Technical Reports Server (NTRS)

    Johnson, Marty E.; Fuller, Chris R.; Jones, Michael G. (Technical Monitor)

    2000-01-01

    In this report both a frequency domain method for creating high level harmonic excitation and a time domain inverse method for creating large pulses in a duct are developed. To create controllable, high level sound an axial array of six JBL-2485 compression drivers was used. The pressure downstream is considered as input voltages to the sources filtered by the natural dynamics of the sources and the duct. It is shown that this dynamic behavior can be compensated for by filtering the inputs such that both time delays and phase changes are taken into account. The methods developed maximize the sound output while (i) keeping within the power constraints of the sources and (ii) maintaining a suitable level of reproduction accuracy. Harmonic excitation pressure levels of over 155dB were created experimentally over a wide frequency range (1000-4000Hz). For pulse excitation there is a tradeoff between accuracy of reproduction and sound level achieved. However, the accurate reproduction of a pulse with a maximum pressure level over 6500Pa was achieved experimentally. It was also shown that the throat connecting the driver to the duct makes it difficult to inject sound just below the cut-on of each acoustic mode (pre cut-on loading effect).

  7. Perceptual assessment of quality of urban soundscapes with combined noise sources and water sounds.

    PubMed

    Jeon, Jin Yong; Lee, Pyoung Jik; You, Jin; Kang, Jian

    2010-03-01

    In this study, urban soundscapes containing combined noise sources were evaluated through field surveys and laboratory experiments. The effect of water sounds on masking urban noises was then examined in order to enhance the soundscape perception. Field surveys in 16 urban spaces were conducted through soundwalking to evaluate the annoyance of combined noise sources. Synthesis curves were derived for the relationships between noise levels and the percentage of highly annoyed (%HA) and the percentage of annoyed (%A) for the combined noise sources. Qualitative analysis was also made using semantic scales for evaluating the quality of the soundscape, and it was shown that the perception of acoustic comfort and loudness was strongly related to the annoyance. A laboratory auditory experiment was then conducted in order to quantify the total annoyance caused by road traffic noise and four types of construction noise. It was shown that the annoyance ratings were related to the types of construction noise in combination with road traffic noise and the level of the road traffic noise. Finally, water sounds were determined to be the best sounds to use for enhancing the urban soundscape. The level of the water sounds should be similar to or not less than 3 dB below the level of the urban noises.

  8. High intensity tone generation by axisymmetric ring cavities on training projectiles

    NASA Technical Reports Server (NTRS)

    Parthasarathy, S. P.; Cho, Y. I.; Back, L. H.

    1984-01-01

    An experimental investigation has been carried out on the production of high intensity tones by axisymmetric ring cavities. Maximum sound production occurs during a double resonance at Strouhal numbers which depend only on the local flow velocity independent of cavity location. Values of sound pressure of about 115 dB at 1 meter distance can be generated by axisymmetric ring cavities on projectiles moving at a relatively low flight speed equal to 65 m/s. Frequencies in the audible range up to several Kilo Hertz can be generated aeroacoustically.

  9. Lymphocytes on sounding rocket flights.

    PubMed

    Cogoli-Greuter, M; Pippia, P; Sciola, L; Cogoli, A

    1994-05-01

    Cell-cell interactions and the formation of cell aggregates are important events in the mitogen-induced lymphocyte activation. The fact that the formation of cell aggregates is only slightly reduced in microgravity suggests that cells are moving and interacting also in space, but direct evidence was still lacking. Here we report on two experiments carried out on a flight of the sounding rocket MAXUS 1B, launched in November 1992 from the base of Esrange in Sweden. The rocket reached the altitude of 716 km and provided 12.5 min of microgravity conditions.

  10. Acoustic Pump

    NASA Technical Reports Server (NTRS)

    Heyman, Joseph S.

    1993-01-01

    Pump uses acoustic-radiation forces. Momentum transferred from sound waves to sound-propagating material in way resulting in net pumping action on material. Acoustic pump is solid-state pump. Requires no moving parts, entirely miniaturized, and does not invade pumped environment. Silent, with no conventional vibration. Used as pump for liquid, suspension, gas, or any other medium interacting with radiation pressure. Also used where solid-state pump needed for reliability and controllability. In microgravity environment, device offers unusual control for low flow rates. For medical or other applications in which contamination cannot be allowed, offers noninvasive pumping force.

  11. Development of a Finite-Difference Time Domain (FDTD) Model for Propagation of Transient Sounds in Very Shallow Water.

    PubMed

    Sprague, Mark W; Luczkovich, Joseph J

    2016-01-01

    This finite-difference time domain (FDTD) model for sound propagation in very shallow water uses pressure and velocity grids with both 3-dimensional Cartesian and 2-dimensional cylindrical implementations. Parameters, including water and sediment properties, can vary in each dimension. Steady-state and transient signals from discrete and distributed sources, such as the surface of a vibrating pile, can be used. The cylindrical implementation uses less computation but requires axial symmetry. The Cartesian implementation allows asymmetry. FDTD calculations compare well with those of a split-step parabolic equation. Applications include modeling the propagation of individual fish sounds, fish aggregation sounds, and distributed sources.

  12. Adaptive near-field beamforming techniques for sound source imaging.

    PubMed

    Cho, Yong Thung; Roan, Michael J

    2009-02-01

    Phased array signal processing techniques such as beamforming have a long history in applications such as sonar for detection and localization of far-field sound sources. Two sometimes competing challenges arise in any type of spatial processing; these are to minimize contributions from directions other than the look direction and minimize the width of the main lobe. To tackle this problem a large body of work has been devoted to the development of adaptive procedures that attempt to minimize side lobe contributions to the spatial processor output. In this paper, two adaptive beamforming procedures-minimum variance distorsionless response and weight optimization to minimize maximum side lobes--are modified for use in source visualization applications to estimate beamforming pressure and intensity using near-field pressure measurements. These adaptive techniques are compared to a fixed near-field focusing technique (both techniques use near-field beamforming weightings focusing at source locations estimated based on spherical wave array manifold vectors with spatial windows). Sound source resolution accuracies of near-field imaging procedures with different weighting strategies are compared using numerical simulations both in anechoic and reverberant environments with random measurement noise. Also, experimental results are given for near-field sound pressure measurements of an enclosed loudspeaker.

  13. Aircraft laser sensing of sound velocity in water - Brillouin scattering

    NASA Technical Reports Server (NTRS)

    Hickman, G. D.; Harding, John M.; Carnes, Michael; Pressman, AL; Kattawar, George W.; Fry, Edward S.

    1991-01-01

    A real-time data source for sound speed in the upper 100 m has been proposed for exploratory development. This data source is planned to be generated via a ship- or aircraft-mounted optical pulsed laser using the spontaneous Brillouin scattering technique. The system should be capable (from a single 10 ns 500 mJ pulse) of yielding range resolved sound speed profiles in water to depths of 75-100 m to an accuracy of 1 m/s. The 100 m profiles will provide the capability of rapidly monitoring the upper-ocean vertical structure. They will also provide an extensive, subsurface-data source for existing real-time, operational ocean nowcast/forecast systems.

  14. Investigation of hydraulic transmission noise sources

    NASA Astrophysics Data System (ADS)

    Klop, Richard J.

    Advanced hydrostatic transmissions and hydraulic hybrids show potential in new market segments such as commercial vehicles and passenger cars. Such new applications regard low noise generation as a high priority, thus, demanding new quiet hydrostatic transmission designs. In this thesis, the aim is to investigate noise sources of hydrostatic transmissions to discover strategies for designing compact and quiet solutions. A model has been developed to capture the interaction of a pump and motor working in a hydrostatic transmission and to predict overall noise sources. This model allows a designer to compare noise sources for various configurations and to design compact and inherently quiet solutions. The model describes dynamics of the system by coupling lumped parameter pump and motor models with a one-dimensional unsteady compressible transmission line model. The model has been verified with dynamic pressure measurements in the line over a wide operating range for several system structures. Simulation studies were performed illustrating sensitivities of several design variables and the potential of the model to design transmissions with minimal noise sources. A semi-anechoic chamber has been designed and constructed suitable for sound intensity measurements that can be used to derive sound power. Measurements proved the potential to reduce audible noise by predicting and reducing both noise sources. Sound power measurements were conducted on a series hybrid transmission test bench to validate the model and compare predicted noise sources with sound power.

  15. On the Possible Detection of Lightning Storms by Elephants

    PubMed Central

    Kelley, Michael C.; Garstang, Michael

    2013-01-01

    Simple Summary We use data similar to that taken by the International Monitoring System for the detection of nuclear explosions, to determine whether elephants might be capable of detecting and locating the source of sounds generated by thunderstorms. Knowledge that elephants might be capable of responding to such storms, particularly at the end of the dry season when migrations are initiated, is of considerable interest to management and conservation. Abstract Theoretical calculations suggest that sounds produced by thunderstorms and detected by a system similar to the International Monitoring System (IMS) for the detection of nuclear explosions at distances ≥100 km, are at sound pressure levels equal to or greater than 6 × 10−3 Pa. Such sound pressure levels are well within the range of elephant hearing. Frequencies carrying these sounds might allow for interaural time delays such that adult elephants could not only hear but could also locate the source of these sounds. Determining whether it is possible for elephants to hear and locate thunderstorms contributes to the question of whether elephant movements are triggered or influenced by these abiotic sounds. PMID:26487406

  16. Numerical flow simulation of a reusable sounding rocket during nose-up rotation

    NASA Astrophysics Data System (ADS)

    Kuzuu, Kazuto; Kitamura, Keiichi; Fujimoto, Keiichiro; Shima, Eiji

    2010-11-01

    Flow around a reusable sounding rocket during nose-up rotation is simulated using unstructured compressible CFD code. While a reusable sounding rocket is expected to reduce the cost of the flight management, it is demanded that this rocket has good performance for wide range of flight conditions from vertical take-off to vertical landing. A rotating body, which corresponds to a vehicle's motion just before vertical landing, is one of flight environments that largely affect its aerodynamic design. Unlike landing of the space shuttle, this vehicle must rotate from gliding position to vertical landing position in nose-up direction. During this rotation, the vehicle generates massive separations in the wake. As a result, induced flow becomes unsteady and could have influence on aerodynamic characteristics of the vehicle. In this study, we focus on the analysis of such dynamic characteristics of the rotating vehicle. An employed numerical code is based on a cell-centered finite volume compressible flow solver applied to a moving grid system. The moving grid is introduced for the analysis of rotating motion. Furthermore, in order to estimate an unsteady turbulence, we employed DDES method as a turbulence model. In this simulation, flight velocity is subsonic. Through this simulation, we discuss the effect on aerodynamic characteristics of a vehicle's shape and motion.

  17. Test of a non-physical barrier consisting of light, sound, and bubble screen to block upstream movement of sea lamprey in an experimental raceway

    USGS Publications Warehouse

    Miehls, Scott M.; Johnson, Nicholas S.; Hrodey, Pete J.

    2017-01-01

    Control of the invasive Sea Lamprey Petromyzon marinus is critical for management of commercial and recreational fisheries in the Laurentian Great Lakes. Use of physical barriers to block Sea Lampreys from spawning habitat is a major component of the control program. However, the resulting interruption of natural streamflow and blockage of nontarget species present substantial challenges. Development of an effective nonphysical barrier would aid the control of Sea Lampreys by eliminating their access to spawning locations while maintaining natural streamflow. We tested the effect of a nonphysical barrier consisting of strobe lights, low-frequency sound, and a bubble screen on the movement of Sea Lampreys in an experimental raceway designed as a two-choice maze with a single main channel fed by two identical inflow channels (one control and one blocked). Sea Lampreys were more likely to move upstream during trials when the strobe light and low-frequency sound were active compared with control trials and trials using the bubble screen alone. For those Sea Lampreys that did move upstream to the confluence of inflow channels, no combination of stimuli or any individual stimulus significantly influenced the likelihood that Sea Lampreys would enter the blocked inflow channel, enter the control channel, or return downstream.

  18. Investigation of the sound generation mechanisms for in-duct orifice plates.

    PubMed

    Tao, Fuyang; Joseph, Phillip; Zhang, Xin; Stalnov, Oksana; Siercke, Matthias; Scheel, Henning

    2017-08-01

    Sound generation due to an orifice plate in a hard-walled flow duct which is commonly used in air distribution systems (ADS) and flow meters is investigated. The aim is to provide an understanding of this noise generation mechanism based on measurements of the source pressure distribution over the orifice plate. A simple model based on Curle's acoustic analogy is described that relates the broadband in-duct sound field to the surface pressure cross spectrum on both sides of the orifice plate. This work describes careful measurements of the surface pressure cross spectrum over the orifice plate from which the surface pressure distribution and correlation length is deduced. This information is then used to predict the radiated in-duct sound field. Agreement within 3 dB between the predicted and directly measured sound fields is obtained, providing direct confirmation that the surface pressure fluctuations acting over the orifice plates are the main noise sources. Based on the developed model, the contributions to the sound field from different radial locations of the orifice plate are calculated. The surface pressure is shown to follow a U 3.9 velocity scaling law and the area over which the surface sources are correlated follows a U 1.8 velocity scaling law.

  19. The Confirmation of the Inverse Square Law Using Diffraction Gratings

    ERIC Educational Resources Information Center

    Papacosta, Pangratios; Linscheid, Nathan

    2014-01-01

    Understanding the inverse square law, how for example the intensity of light or sound varies with distance, presents conceptual and mathematical challenges. Students know intuitively that intensity decreases with distance. A light source appears dimmer and sound gets fainter as the distance from the source increases. The difficulty is in…

  20. Gaussian impurity moving through a Bose-Einstein superfluid

    NASA Astrophysics Data System (ADS)

    Pinsker, Florian

    2017-09-01

    In this paper a finite Gaussian impurity moving through an equilibrium Bose-Einstein condensate at T = 0 is studied. The problem can be described by a Gross-Pitaevskii equation, which is solved perturbatively. The analysis is done for systems of 2 and 3 spatial dimensions. The Bogoliubov equation solutions for the condensate perturbed by a finite impurity are calculated in the co-moving frame. From these solutions the total energy of the perturbed system is determined as a function of the width and the amplitude of the moving Gaussian impurity and its velocity. In addition we derive the drag force the finite sized impurity approximately experiences as it moves through the superfluid, which proves the existence of a superfluid phase for finite extensions of the impurities below the speed of sound. Finally we find that the force increases with velocity until an inflection point from which it decreases again in 2 and 3d.

  1. The sound strength parameter G and its importance in evaluating and planning the acoustics of halls for music.

    PubMed

    Beranek, Leo

    2011-05-01

    The parameter, "Strength of Sound G" is closely related to loudness. Its magnitude is dependent, inversely, on the total sound absorption in a room. By comparison, the reverberation time (RT) is both inversely related to the total sound absorption in a hall and directly related to its cubic volume. Hence, G and RT in combination are vital in planning the acoustics of a concert hall. A newly proposed "Bass Index" is related to the loudness of the bass sound and equals the value of G at 125 Hz in decibels minus its value at mid-frequencies. Listener envelopment (LEV) is shown for most halls to be directly related to the mid-frequency value of G. The broadening of sound, i.e., apparent source width (ASW) is given by degree of source broadening (DSB) which is determined from the combined effect of early lateral reflections as measured by binaural quality index (BQI) and strength G. The optimum values and limits of these parameters are discussed.

  2. The role of long-term familiarity and attentional maintenance in short-term memory for timbre.

    PubMed

    Siedenburg, Kai; McAdams, Stephen

    2017-04-01

    We study short-term recognition of timbre using familiar recorded tones from acoustic instruments and unfamiliar transformed tones that do not readily evoke sound-source categories. Participants indicated whether the timbre of a probe sound matched with one of three previously presented sounds (item recognition). In Exp. 1, musicians better recognised familiar acoustic compared to unfamiliar synthetic sounds, and this advantage was particularly large in the medial serial position. There was a strong correlation between correct rejection rate and the mean perceptual dissimilarity of the probe to the tones from the sequence. Exp. 2 compared musicians' and non-musicians' performance with concurrent articulatory suppression, visual interference, and with a silent control condition. Both suppression tasks disrupted performance by a similar margin, regardless of musical training of participants or type of sounds. Our results suggest that familiarity with sound source categories and attention play important roles in short-term memory for timbre, which rules out accounts solely based on sensory persistence.

  3. Octopus Cells in the Posteroventral Cochlear Nucleus Provide the Main Excitatory Input to the Superior Paraolivary Nucleus

    PubMed Central

    Felix II, Richard A.; Gourévitch, Boris; Gómez-Álvarez, Marcelo; Leijon, Sara C. M.; Saldaña, Enrique; Magnusson, Anna K.

    2017-01-01

    Auditory streaming enables perception and interpretation of complex acoustic environments that contain competing sound sources. At early stages of central processing, sounds are segregated into separate streams representing attributes that later merge into acoustic objects. Streaming of temporal cues is critical for perceiving vocal communication, such as human speech, but our understanding of circuits that underlie this process is lacking, particularly at subcortical levels. The superior paraolivary nucleus (SPON), a prominent group of inhibitory neurons in the mammalian brainstem, has been implicated in processing temporal information needed for the segmentation of ongoing complex sounds into discrete events. The SPON requires temporally precise and robust excitatory input(s) to convey information about the steep rise in sound amplitude that marks the onset of voiced sound elements. Unfortunately, the sources of excitation to the SPON and the impact of these inputs on the behavior of SPON neurons have yet to be resolved. Using anatomical tract tracing and immunohistochemistry, we identified octopus cells in the contralateral cochlear nucleus (CN) as the primary source of excitatory input to the SPON. Cluster analysis of miniature excitatory events also indicated that the majority of SPON neurons receive one type of excitatory input. Precise octopus cell-driven onset spiking coupled with transient offset spiking make SPON responses well-suited to signal transitions in sound energy contained in vocalizations. Targets of octopus cell projections, including the SPON, are strongly implicated in the processing of temporal sound features, which suggests a common pathway that conveys information critical for perception of complex natural sounds. PMID:28620283

  4. Harmonic Hopping, and Both Punctuated and Gradual Evolution of Acoustic Characters in Selasphorus Hummingbird Tail-Feathers

    PubMed Central

    Clark, Christopher James

    2014-01-01

    Models of character evolution often assume a single mode of evolutionary change, such as continuous, or discrete. Here I provide an example in which a character exhibits both types of change. Hummingbirds in the genus Selasphorus produce sound with fluttering tail-feathers during courtship. The ancestral character state within Selasphorus is production of sound with an inner tail-feather, R2, in which the sound usually evolves gradually. Calliope and Allen's Hummingbirds have evolved autapomorphic acoustic mechanisms that involve feather-feather interactions. I develop a source-filter model of these interactions. The ‘source’ comprises feather(s) that are both necessary and sufficient for sound production, and are aerodynamically coupled to neighboring feathers, which act as filters. Filters are unnecessary or insufficient for sound production, but may evolve to become sources. Allen's Hummingbird has evolved to produce sound with two sources, one with feather R3, another frequency-modulated sound with R4, and their interaction frequencies. Allen's R2 retains the ancestral character state, a ∼1 kHz “ghost” fundamental frequency masked by R3, which is revealed when R3 is experimentally removed. In the ancestor to Allen's Hummingbird, the dominant frequency has ‘hopped’ to the second harmonic without passing through intermediate frequencies. This demonstrates that although the fundamental frequency of a communication sound may usually evolve gradually, occasional jumps from one character state to another can occur in a discrete fashion. Accordingly, mapping acoustic characters on a phylogeny may produce misleading results if the physical mechanism of production is not known. PMID:24722049

  5. 36 CFR 251.51 - Definitions.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... of motion picture, videotaping, sound recording, or any other moving image or audio recording... category—A processing or monitoring category requiring more than 50 hours of agency time to process an application for a special use authorization (processing category 6 and, in certain situations, processing...

  6. 36 CFR 251.51 - Definitions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... of motion picture, videotaping, sound recording, or any other moving image or audio recording.... Major category—A processing or monitoring category requiring more than 50 hours of agency time to process an application for a special use authorization (processing category 6 and, in certain situations...

  7. 36 CFR 251.51 - Definitions.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... of motion picture, videotaping, sound recording, or any other moving image or audio recording... category—A processing or monitoring category requiring more than 50 hours of agency time to process an application for a special use authorization (processing category 6 and, in certain situations, processing...

  8. 36 CFR 251.51 - Definitions.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... of motion picture, videotaping, sound recording, or any other moving image or audio recording... category—A processing or monitoring category requiring more than 50 hours of agency time to process an application for a special use authorization (processing category 6 and, in certain situations, processing...

  9. Augmented awareness

    USDA-ARS?s Scientific Manuscript database

    The sounds insects make as they move and feed can be amplified 100–1000x by vibration sensor systems, which makes them easily detectable with headphones. In 2008, I visited Mackay and Bundaberg in Queensland to search for greyback canegrub (Dermolepida albohirtum) and other insect pests in sugarcane...

  10. Addressing Measurement Issues Related to Bullying Involvement

    ERIC Educational Resources Information Center

    Casper, Deborah M.; Meter, Diana J.; Card, Noel A.

    2015-01-01

    In this article, we address measurement issues related to select aspects of bullying involvement with the goal of moving psychometrically sound measurement practices toward applied bullying research. We first provide a nontechnical introduction to psychometric considerations in measuring bullying involvement, highlighting the importance of…

  11. A Green Soundscape Index (GSI): The potential of assessing the perceived balance between natural sound and traffic noise.

    PubMed

    Kogan, Pablo; Arenas, Jorge P; Bermejo, Fernando; Hinalaf, María; Turra, Bruno

    2018-06-13

    Urban soundscapes are dynamic and complex multivariable environmental systems. Soundscapes can be organized into three main entities containing the multiple variables: Experienced Environment (EE), Acoustic Environment (AE), and Extra-Acoustic Environment (XE). This work applies a multidimensional and synchronic data-collecting methodology at eight urban environments in the city of Córdoba, Argentina. The EE was assessed by means of surveys, the AE by acoustic measurements and audio recordings, and the XE by photos, video, and complementary sources. In total, 39 measurement locations were considered, where data corresponding to 61 AE and 203 EE were collected. Multivariate analysis and GIS techniques were used for data processing. The types of sound sources perceived, and their extents make up part of the collected variables that belong to the EE, i.e. traffic, people, natural sounds, and others. Sources explaining most of the variance were traffic noise and natural sounds. Thus, a Green Soundscape Index (GSI) is defined here as the ratio of the perceived extents of natural sounds to traffic noise. Collected data were divided into three ranges according to GSI value: 1) perceptual predominance of traffic noise, 2) balanced perception, and 3) perceptual predominance of natural sounds. For each group, three additional variables from the EE and three from the AE were applied, which reported significant differences, especially between ranges 1 and 2 with 3. These results confirm the key role of perceiving natural sounds in a town environment and also support the proposal of a GSI as a valuable indicator to classify urban soundscapes. In addition, the collected GSI-related data significantly helps to assess the overall soundscape. It is noted that this proposed simple perceptual index not only allows one to assess and classify urban soundscapes but also contributes greatly toward a technique for separating environmental sound sources. Copyright © 2018 Elsevier B.V. All rights reserved.

  12. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Efroymson, Rebecca Ann; Hargrove, William Walter; Suter, Glenn

    A multi-stressor risk assessment was conducted at Yuma Proving Ground, Arizona, as a demonstration of the Military Ecological Risk Assessment Framework. The focus of the assessment was a testing program at Cibola Range, which involved an Apache Longbow helicopter firing Hellfire missiles at moving targets, M60-A1 tanks. This paper focuses on the wildlife risk assessment for the helicopter overflight. The primary stressors were sound and the view of the aircraft. Exposure to desert mule deer (Odocoileus hemionus crooki) was quantified using Air Force sound contour programs NOISEMAP and MR_NMAP, which gave very different results. Slant distance from helicopters to deermore » was also used as a measure of exposure that integrated risk from sound and view of the aircraft. Exposure-response models for the characterization of effects consisted of behavioral thresholds in sound exposure level or maximum sound level units or slant distance. Available sound thresholds were limited for desert mule deer, but a distribution of slant-distance thresholds was available for ungulates. The risk characterization used a weight-of-evidence approach and concluded that risk to mule deer behavior from the Apache overflight is uncertain, but that no risk to mule deer abundance and reproduction is expected.« less

  13. Sound produced by an oscillating arc in a high-pressure gas

    NASA Astrophysics Data System (ADS)

    Popov, Fedor K.; Shneider, Mikhail N.

    2017-08-01

    We suggest a simple theory to describe the sound generated by small periodic perturbations of a cylindrical arc in a dense gas. Theoretical analysis was done within the framework of the non-self-consistent channel arc model and supplemented with time-dependent gas dynamic equations. It is shown that an arc with power amplitude oscillations on the order of several percent is a source of sound whose intensity is comparable with external ultrasound sources used in experiments to increase the yield of nanoparticles in the high pressure arc systems for nanoparticle synthesis.

  14. Focusing and directional beaming effects of airborne sound through a planar lens with zigzag slits

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tang, Kun; Qiu, Chunyin, E-mail: cyqiu@whu.edu.cn; Lu, Jiuyang

    2015-01-14

    Based on the Huygens-Fresnel principle, we design a planar lens to efficiently realize the interconversion between the point-like sound source and Gaussian beam in ambient air. The lens is constructed by a planar plate perforated elaborately with a nonuniform array of zigzag slits, where the slit exits act as subwavelength-sized secondary sources carrying desired sound responses. The experiments operated at audible regime agree well with the theoretical predictions. This compact device could be useful in daily life applications, such as for medical and detection purposes.

  15. High-frequency monopole sound source for anechoic chamber qualification

    NASA Astrophysics Data System (ADS)

    Saussus, Patrick; Cunefare, Kenneth A.

    2003-04-01

    Anechoic chamber qualification procedures require the use of an omnidirectional monopole sound source. Required characteristics for these monopole sources are explicitly listed in ISO 3745. Building a high-frequency monopole source that meets these characteristics has proved difficult due to the size limitations imposed by small wavelengths at high frequency. A prototype design developed for use in hemianechoic chambers employs telescoping tubes, which act as an inverse horn. This same design can be used in anechoic chambers, with minor adaptations. A series of gradually decreasing brass telescoping tubes is attached to the throat of a well-insulated high-frequency compression driver. Therefore, all of the sound emitted from the driver travels through the horn and exits through an opening of approximately 2.5 mm. Directivity test data show that this design meets all of the requirements set forth by ISO 3745.

  16. Second-sound studies of coflow and counterflow of superfluid {sup 4}He in channels

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Varga, Emil; Skrbek, L.; Babuin, Simone, E-mail: babuin@fzu.cz

    2015-06-15

    We report a comprehensive study of turbulent superfluid {sup 4}He flow through a channel of square cross section. We study for the first time two distinct flow configurations with the same apparatus: coflow (normal and superfluid components move in the same direction), and counterflow (normal and superfluid components move in opposite directions). We realise also a variation of counterflow with the same relative velocity, but where the superfluid component moves while there is no net flow of the normal component through the channel, i.e., pure superflow. We use the second-sound attenuation technique to measure the density of quantised vortex linesmore » in the temperature range 1.2 K ≲ T ≲ T{sub λ} ≈ 2.18 K and for flow velocities from about 1 mm/s up to almost 1 m/s in fully developed turbulence. We find that both the steady-state and temporal decay of the turbulence significantly differ in the three flow configurations, yielding an interesting insight into two-fluid hydrodynamics. In both pure superflow and counterflow, the same scaling of vortex line density with counterflow velocity is observed, L∝V{sub cf}{sup 2}, with a pronounced temperature dependence; in coflow instead, the vortex line density scales with velocity as L ∝ V{sup 3/2} and is temperature independent; we provide theoretical explanations for these observations. Further, we develop a new promising technique to use different second-sound resonant modes to probe the spatial distribution of quantised vortices in the direction perpendicular to the flow. Preliminary measurements indicate that coflow is less homogeneous than counterflow/superflow, with a denser concentration of vortices between the centre of the channel and its walls.« less

  17. Assessment of sound levels in a neonatal intensive care unit in tabriz, iran.

    PubMed

    Valizadeh, Sousan; Bagher Hosseini, Mohammad; Alavi, Nasrinsadat; Asadollahi, Malihe; Kashefimehr, Siamak

    2013-03-01

    High levels of sound have several negative effects, such as noise-induced hearing loss and delayed growth and development, on premature infants in neonatal intensive care units (NICUs). In order to reduce sound levels, they should first be measured. This study was performed to assess sound levels and determine sources of noise in the NICU of Alzahra Teaching Hospital (Tabriz, Iran). In a descriptive study, 24 hours in 4 workdays were randomly selected. Equivalent continuous sound level (Leq), sound level that is exceeded only 10% of the time (L10), maximum sound level (Lmax), and peak instantaneous sound pressure level (Lzpeak) were measured by CEL-440 sound level meter (SLM) at 6 fixed locations in the NICU. Data was collected using a questionnaire. SPSS13 was then used for data analysis. Mean values of Leq, L10, and Lmax were determined as 63.46 dBA, 65.81 dBA, and 71.30 dBA, respectively. They were all higher than standard levels (Leq < 45 dB, L10 ≤50 dB, and Lmax ≤65 dB). The highest Leq was measured at the time of nurse rounds. Leq was directly correlated with the number of staff members present in the ward. Finally, sources of noise were ordered based on their intensity. Considering that sound levels were higher than standard levels in our studied NICU, it is necessary to adopt policies to reduce sound.

  18. Assessment of Sound Levels in a Neonatal Intensive Care Unit in Tabriz, Iran

    PubMed Central

    Valizadeh, Sousan; Bagher Hosseini, Mohammad; Alavi, Nasrinsadat; Asadollahi, Malihe; Kashefimehr, Siamak

    2013-01-01

    Introduction: High levels of sound have several negative effects, such as noise-induced hearing loss and delayed growth and development, on premature infants in neonatal intensive care units (NICUs). In order to reduce sound levels, they should first be measured. This study was performed to assess sound levels and determine sources of noise in the NICU of Alzahra Teaching Hospital (Tabriz, Iran). Methods: In a descriptive study, 24 hours in 4 workdays were randomly selected. Equivalent continuous sound level (Leq), sound level that is exceeded only 10% of the time (L10), maximum sound level (Lmax), and peak instantaneous sound pressure level (Lzpeak) were measured by CEL-440 sound level meter (SLM) at 6 fixed locations in the NICU. Data was collected using a questionnaire. SPSS13 was then used for data analysis. Results: Mean values of Leq, L10, and Lmax were determined as 63.46 dBA, 65.81 dBA, and 71.30 dBA, respectively. They were all higher than standard levels (Leq < 45 dB, L10 ≤50 dB, and Lmax ≤65 dB). The highest Leq was measured at the time of nurse rounds. Leq was directly correlated with the number of staff members present in the ward. Finally, sources of noise were ordered based on their intensity. Conclusion: Considering that sound levels were higher than standard levels in our studied NICU, it is necessary to adopt policies to reduce sound. PMID:25276706

  19. Are high perches in the blackcap Sylvia atricapilla song or listening posts? A sound transmission study.

    PubMed

    Mathevon, Nicolas; Dabelsteen, Torben; Blumenrath, Sandra H

    2005-01-01

    Birds often sing from high perches referred to as song posts. However, birds also listen and keep a lookout from these perches. We used a sound transmission experiment to investigate the changes for receiving and sending conditions that a territorial songbird may experience by moving upwards in the vegetation. Representative song elements of the blackcap Sylvia atricapilla were transmitted in a forest habitat in spring using a complete factorial design with natural transmission distances and speaker and microphone heights. Four aspects of sound degradation were quantified: signal-to-noise ratio, excess attenuation, distortion within the sounds determined as a blur ratio, and prolongation of the sounds with "tails" of echoes determined as a tail-to-signal ratio. All four measures indicated that degradation decreased with speaker and microphone height. However, the decrease was considerably higher for the microphone than for the speaker. This suggests that choosing high perches in a forest at spring results in more benefits to blackcaps in terms of improved communication conditions when they act as receivers than as senders.

  20. Effects of the Exxon Valdez oil spill on the abundance and distribution of humpback whales (`megaptera novaeangliae`) in Prince William Sound. Marine mammal study number 1. Exxon Valdez oil spill state/federal natural resource damage assessment final report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Dahlheim, M.E.; Von Ziegesar, O.

    1993-12-01

    Photo-identification studies of Prince William Sound humpback whales were conducted from May to September in 1989 and 1990 to assess the impact on the spill on humpback whale life history and ecology. In 1989, concurrent studies were conducted in Southeast Alaska on humpback whales to determine if whales avoided contaminated waters of Prince William Sound and moved to other northern feeding areas. In 1989, photographic analysis of Prince William Sound humpbacks resulted in the identification of 59 whales. In 1990, 66 whales were documented. The increase in whale sightings may have been due to the increase in effort during themore » 1989 and 1990 season. Because of the difference in survey effort before and after the spill, it is impossible to determine if there was a difference in the number of humpback whales using the Sound. Distribution varied among years and may be related to prey distribution.« less

  1. The Processing of Attended and Predicted Sounds in Time.

    PubMed

    Paris, Tim; Kim, Jeesun; Davis, Chris

    2016-01-01

    Neural responses to an attended event are typically enhanced relative to those from an unattended one (attention enhancement). Conversely, neural responses to a predicted event are typically reduced relative to those from an unpredicted one (prediction suppression). What remains to be established is what happens with attended and predicted events. To examine the interaction between attention and prediction, we combined two robust paradigms developed for studying attention and prediction effects on ERPs into an orthogonal design. Participants were presented with sounds in attended or unattended intervals with onsets that were either predicted by a moving visual cue or unpredicted (no cue was provided). We demonstrated an N1 enhancement effect for attended sounds and an N1 suppression effect for predicted sounds; furthermore, an interaction between these effects was found that emerged early in the N1 (50-95 msec), indicating that attention enhancement only occurred when the sound was unpredicted. This pattern of results can be explained by the precision of the predictive cue that reduces the need for attention selection in the attended and predicted condition.

  2. Electrophysiological correlates of cocktail-party listening.

    PubMed

    Lewald, Jörg; Getzmann, Stephan

    2015-10-01

    Detecting, localizing, and selectively attending to a particular sound source of interest in complex auditory scenes composed of multiple competing sources is a remarkable capacity of the human auditory system. The neural basis of this so-called "cocktail-party effect" has remained largely unknown. Here, we studied the cortical network engaged in solving the "cocktail-party" problem, using event-related potentials (ERPs) in combination with two tasks demanding horizontal localization of a naturalistic target sound presented either in silence or in the presence of multiple competing sound sources. Presentation of multiple sound sources, as compared to single sources, induced an increased P1 amplitude, a reduction in N1, and a strong N2 component, resulting in a pronounced negativity in the ERP difference waveform (N2d) around 260 ms after stimulus onset. About 100 ms later, the anterior contralateral N2 subcomponent (N2ac) occurred in the multiple-sources condition, as computed from the amplitude difference for targets in the left minus right hemispaces. Cortical source analyses of the ERP modulation, resulting from the contrast of multiple vs. single sources, generally revealed an initial enhancement of electrical activity in right temporo-parietal areas, including auditory cortex, by multiple sources (at P1) that is followed by a reduction, with the primary sources shifting from right inferior parietal lobule (at N1) to left dorso-frontal cortex (at N2d). Thus, cocktail-party listening, as compared to single-source localization, appears to be based on a complex chronology of successive electrical activities within a specific cortical network involved in spatial hearing in complex situations. Copyright © 2015 Elsevier B.V. All rights reserved.

  3. Determination of equivalent sound speed profiles for ray tracing in near-ground sound propagation.

    PubMed

    Prospathopoulos, John M; Voutsinas, Spyros G

    2007-09-01

    The determination of appropriate sound speed profiles in the modeling of near-ground propagation using a ray tracing method is investigated using a ray tracing model which is capable of performing axisymmetric calculations of the sound field around an isolated source. Eigenrays are traced using an iterative procedure which integrates the trajectory equations for each ray launched from the source at a specific direction. The calculation of sound energy losses is made by introducing appropriate coefficients to the equations representing the effect of ground and atmospheric absorption and the interaction with the atmospheric turbulence. The model is validated against analytical and numerical predictions of other methodologies for simple cases, as well as against measurements for nonrefractive atmospheric environments. A systematic investigation for near-ground propagation in downward and upward refractive atmosphere is made using experimental data. Guidelines for the suitable simulation of the wind velocity profile are derived by correlating predictions with measurements.

  4. Design and test of the 172K fluidic rudder

    NASA Technical Reports Server (NTRS)

    Belsterling, C. A.

    1978-01-01

    Progress in the development of concepts for control of aircraft without moving parts or a separate source of power is described. The design and wind tunnel tests of a full scale fluidic rudder for a Cessna 172K aircraft, intended for subsequent flight tests were documented. The 172K fluidic rudder was designed to provide a control force equivalent to 3.3 degrees of deflection of the conventional rudder. In spite of an extremely thin airfoil, cascaded fluidic amplifiers were built to fit, with the capacity for generating the required level of control force. Wind tunnel tests demonstrated that the principles of lift control using ram air power are sound and reliable under all flight conditions. The tests also demonstrated that the performance of the 172K fluidic rudder is not acceptable for flight tests until the design of the scoop is modified to prevent interference with the lift control phenomenon.

  5. Acoustic centering of sources measured by surrounding spherical microphone arrays.

    PubMed

    Hagai, Ilan Ben; Pollow, Martin; Vorländer, Michael; Rafaely, Boaz

    2011-10-01

    The radiation patterns of acoustic sources have great significance in a wide range of applications, such as measuring the directivity of loudspeakers and investigating the radiation of musical instruments for auralization. Recently, surrounding spherical microphone arrays have been studied for sound field analysis, facilitating measurement of the pressure around a sphere and the computation of the spherical harmonics spectrum of the sound source. However, the sound radiation pattern may be affected by the location of the source inside the microphone array, which is an undesirable property when aiming to characterize source radiation in a unique manner. This paper presents a theoretical analysis of the spherical harmonics spectrum of spatially translated sources and defines four measures for the misalignment of the acoustic center of a radiating source. Optimization is used to promote optimal alignment based on the proposed measures and the errors caused by numerical and array-order limitations are investigated. This methodology is examined using both simulated and experimental data in order to investigate the performance and limitations of the different alignment methods. © 2011 Acoustical Society of America

  6. Propagation characteristics of audible noise generated by single corona source under positive DC voltage

    NASA Astrophysics Data System (ADS)

    Li, Xuebao; Cui, Xiang; Lu, Tiebing; Wang, Donglai

    2017-10-01

    The directivity and lateral profile of corona-generated audible noise (AN) from a single corona source are measured through experiments carried out in the semi-anechoic laboratory. The experimental results show that the waveform of corona-generated AN consists of a series of random sound pressure pulses whose pulse amplitudes decrease with the increase of measurement distance. A single corona source can be regarded as a non-directional AN source, and the A-weighted SPL (sound pressure level) decreases 6 dB(A) as doubling the measurement distance. Then, qualitative explanations for the rationality of treating the single corona source as a point source are given on the basis of the Ingard's theory for sound generation in corona discharge. Furthermore, we take into consideration of the ground reflection and the air attenuation to reconstruct the propagation features of AN from the single corona source. The calculated results agree with the measurement well, which validates the propagation model. Finally, the influence of the ground reflection on the SPL is presented in the paper.

  7. Cognitive and Linguistic Sources of Variance in 2-Year-Olds' Speech-Sound Discrimination: A Preliminary Investigation

    ERIC Educational Resources Information Center

    Lalonde, Kaylah; Holt, Rachael Frush

    2014-01-01

    Purpose: This preliminary investigation explored potential cognitive and linguistic sources of variance in 2- year-olds' speech-sound discrimination by using the toddler change/no-change procedure and examined whether modifications would result in a procedure that can be used consistently with younger 2-year-olds. Method: Twenty typically…

  8. The use of an intraoral electrolarynx for an edentulous patient: a clinical report.

    PubMed

    Wee, Alvin G; Wee, Lisa A; Cheng, Ansgar C; Cwynar, Roger B

    2004-06-01

    This clinical report describes the clinical requirements, treatment sequence, and use of a relatively new intraoral electrolarynx for a completely edentulous patient. This device consists of a sound source attached to the maxilla and a hand-held controller unit that controls the pitch and volume of the intraoral sound source via transmitted radio waves.

  9. Spherical beamforming for spherical array with impedance surface

    NASA Astrophysics Data System (ADS)

    Tontiwattanakul, Khemapat

    2018-01-01

    Spherical microphone array beamforming has been a popular research topic for recent years. Due to their isotropic beam in three dimensional spaces as well as a certain frequency range, the arrays are widely used in many applications such as sound field recording, acoustic beamforming, and noise source localisation. The body of a spherical array is usually considered perfectly rigid. A sound field captured by the sensors on spherical array can be decomposed into a series of spherical harmonics. In noise source localisation, the amplitude density of sound sources is estimated and illustrated by mean of colour maps. In this work, a rigid spherical array covered by fibrous materials is studied via numerical simulation and the performance of the spherical beamforming is discussed.

  10. Study on the Non-contact Acoustic Inspection Method for Concrete Structures by using Strong Ultrasonic Sound source

    NASA Astrophysics Data System (ADS)

    Sugimoto, Tsuneyoshi; Uechi, Itsuki; Sugimoto, Kazuko; Utagawa, Noriyuki; Katakura, Kageyoshi

    Hammering test is widely used to inspect the defects in concrete structures. However, this method has a major difficulty in inspect at high-places, such as a tunnel ceiling or a bridge girder. Moreover, its detection accuracy is dependent on a tester's experience. Therefore, we study about the non-contact acoustic inspection method of the concrete structure using the air borne sound wave and a laser Doppler vibrometer. In this method, the concrete surface is excited by air-borne sound wave emitted with a long range acoustic device (LRAD), and the vibration velocity on the concrete surface is measured by a laser Doppler vibrometer. A defect part is detected by the same flexural resonance as the hammer method. It is already shown clearly that detection of a defect can be performed from a long distance of 5 m or more using a concrete test object. Moreover, it is shown that a real concrete structure can also be applied. However, when the conventional LRAD was used as a sound source, there were problems, such as restrictions of a measurement angle and the surrounding noise. In order to solve these problems, basic examination which used the strong ultrasonic wave sound source was carried out. In the experiment, the concrete test object which includes an imitation defect from 5-m distance was used. From the experimental result, when the ultrasonic sound source was used, restrictions of a measurement angle become less severe and it was shown that circumference noise also falls dramatically.

  11. Measurement of the speed of sound by observation of the Mach cones in a complex plasma under microgravity conditions

    NASA Astrophysics Data System (ADS)

    Zhukhovitskii, D. I.; Fortov, V. E.; Molotkov, V. I.; Lipaev, A. M.; Naumkin, V. N.; Thomas, H. M.; Ivlev, A. V.; Schwabe, M.; Morfill, G. E.

    2015-02-01

    We report the first observation of the Mach cones excited by a larger microparticle (projectile) moving through a cloud of smaller microparticles (dust) in a complex plasma with neon as a buffer gas under microgravity conditions. A collective motion of the dust particles occurs as propagation of the contact discontinuity. The corresponding speed of sound was measured by a special method of the Mach cone visualization. The measurement results are incompatible with the theory of ion acoustic waves. The estimate for the pressure in a strongly coupled Coulomb system and a scaling law for the complex plasma make it possible to derive an evaluation for the speed of sound, which is in a reasonable agreement with the experiments in complex plasmas.

  12. Mechanics and materials in middle ear reconstruction.

    PubMed

    Lesser, T H; Williams, K R; Blayney, A W

    1991-02-01

    The normal anatomy and physiology of the middle ear is not reproduced in ossiculoplasty and an artificial mechanism for the transmission of sound results. This is true for all types of graft, be they of natural or man-made material. There are, therefore, 2 areas for consideration when looking at the problems encountered in such reconstructions: first, the materials' biocompatability and, secondly, the mechanical effects of the positioning of the graft in the reconstructed ossicular chain. The present work examines these mechanical effects using the finite element method to determine stress and displacement levels in the reconstructed ossicular chain. It is found that the stress levels at the implant-stapedial joint increases as the implant is gradually moved down the malleus. In contrast there is thought to be an increase in sound transmission as the implant is moved down the malleus. Changes in rigidity and hardness of the implant appear to make only modest stress attenuations at the implant-stapes interface.

  13. An Experimental Investigation of the Flow Structure of Supersonic Impinging Jets

    NASA Technical Reports Server (NTRS)

    Henderson, Brenda; Bridges, James; Wernet, Mark

    2002-01-01

    An experimental investigation into the jet structure associated with sound production by a supersonic impinging jet is presented. Large plate impinging tones are investigated for a nozzle pressure ratio (NPR) of 4 and nozzle-to-plate spacings between 1 and 5 nozzle exit diameters, where NPR is equal to the ratio of the stagnation pressure to the pressure at the nozzle lip. Results from phase-locked shadowgraph and phase-averaged digital particle image velocimetry (DPIV) studies indicate that, during the oscillation cycle, the Mach disk oscillates axially, a well defined recirculation zone is created in the subsonic impingement region and moves toward the plate, and the compression and expansion regions in the outer supersonic flow move downstream, Sound appears to be generated in the wall jet at approximately 2.6R from the jet axis, where R is the nozzle exit radius. The oscillatory motion in the wall jet is the result of the periodic fluid motion in the near wall region.

  14. Accuracy Study of the Space-Time CE/SE Method for Computational Aeroacoustics Problems Involving Shock Waves

    NASA Technical Reports Server (NTRS)

    Wang, Xiao Yen; Chang, Sin-Chung; Jorgenson, Philip C. E.

    1999-01-01

    The space-time conservation element and solution element(CE/SE) method is used to study the sound-shock interaction problem. The order of accuracy of numerical schemes is investigated. The linear model problem.govemed by the 1-D scalar convection equation, sound-shock interaction problem governed by the 1-D Euler equations, and the 1-D shock-tube problem which involves moving shock waves and contact surfaces are solved to investigate the order of accuracy of numerical schemes. It is concluded that the accuracy of the CE/SE numerical scheme with designed 2nd-order accuracy becomes 1st order when a moving shock wave exists. However, the absolute error in the CE/SE solution downstream of the shock wave is on the same order as that obtained using a fourth-order accurate essentially nonoscillatory (ENO) scheme. No special techniques are used for either high-frequency low-amplitude waves or shock waves.

  15. 75 FR 39915 - Marine Mammals; File No. 15483

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-07-13

    ... whales adjust their bearing to avoid received sound pressure levels greater than 120 dB, which would... marine mammals may be taken by Level B harassment as researchers attempt to provoke an avoidance response through sound transmission into their environment. The sound source consists of a transmitter and...

  16. 24 CFR 51.103 - Criteria and standards.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ...-night average sound level produced as the result of the accumulation of noise from all sources contributing to the external noise environment at the site. Day-night average sound level, abbreviated as DNL and symbolized as Ldn, is the 24-hour average sound level, in decibels, obtained after addition of 10...

  17. The origin of Korotkoff sounds and the accuracy of auscultatory blood pressure measurements.

    PubMed

    Babbs, Charles F

    2015-12-01

    This study explores the hypothesis that the sharper, high frequency Korotkoff sounds come from resonant motion of the arterial wall, which begins after the artery transitions from a buckled state to an expanding state. The motions of one mass, two nonlinear springs, and one damper, driven by transmural pressure under the cuff, are used to model and compute the Korotkoff sounds according to principles of classical Newtonian physics. The natural resonance of this spring-mass-damper system provides a concise, yet rigorous, explanation for the origin of Korotkoff sounds. Fundamentally, wall stretching in expansion requires more force than wall bending in buckling. At cuff pressures between systolic and diastolic arterial pressure, audible vibrations (> 40 Hz) occur during early expansion of the artery wall beyond its zero pressure radius after the outward moving mass of tissue experiences sudden deceleration, caused by the discontinuity in stiffness between bucked and expanded states. The idealized spring-mass-damper model faithfully reproduces the time-domain waveforms of actual Korotkoff sounds in humans. Appearance of arterial sounds occurs at or just above the level of systolic pressure. Disappearance of arterial sounds occurs at or just above the level of diastolic pressure. Muffling of the sounds is explained by increased resistance of the artery to collapse, caused by downstream venous engorgement. A simple analytical model can define the physical origin of Korotkoff sounds, suggesting improved mechanical or electronic filters for their selective detection and confirming the disappearance of the Korotkoff sounds as the optimal diastolic end point. Copyright © 2015 American Society of Hypertension. Published by Elsevier Inc. All rights reserved.

  18. Characterisation of structure-borne sound source using reception plate method.

    PubMed

    Putra, A; Saari, N F; Bakri, H; Ramlan, R; Dan, R M

    2013-01-01

    A laboratory-based experiment procedure of reception plate method for structure-borne sound source characterisation is reported in this paper. The method uses the assumption that the input power from the source installed on the plate is equal to the power dissipated by the plate. In this experiment, rectangular plates having high and low mobility relative to that of the source were used as the reception plates and a small electric fan motor was acting as the structure-borne source. The data representing the source characteristics, namely, the free velocity and the source mobility, were obtained and compared with those from direct measurement. Assumptions and constraints employing this method are discussed.

  19. Complete data listings for CSEM soundings on Kilauea Volcano, Hawaii

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kauahikaua, J.; Jackson, D.B.; Zablocki, C.J.

    1983-01-01

    This document contains complete data from a controlled-source electromagnetic (CSEM) sounding/mapping project at Kilauea volcano, Hawaii. The data were obtained at 46 locations about a fixed-location, horizontal, polygonal loop source in the summit area of the volcano. The data consist of magnetic field amplitudes and phases at excitation frequencies between 0.04 and 8 Hz. The vector components were measured in a cylindrical coordinate system centered on the loop source. 5 references.

  20. The Physiological Basis of Chinese Höömii Generation.

    PubMed

    Li, Gelin; Hou, Qian

    2017-01-01

    The study aimed to investigate the physiological basis of vibration mode of sound source of a variety of Mongolian höömii forms of singing in China. The participant is a Mongolian höömii performing artist who was recommended by the Chinese Medical Association of Art. He used three types of höömii, namely vibration höömii, whistle höömii, and overtone höömii, which were compared with general comfortable pronunciation of /i:/ as control. Phonation was observed during /i:/. A laryngostroboscope (Storz) was used to determine vibration source-mucosal wave in the throat. For vibration höömii, bilateral ventricular folds approximated to the midline and made contact at the midline during pronunciation. Ventricular and vocal folds oscillated together as a single unit to form a composite vibration (double oscillator) sound source. For whistle höömii, ventricular folds approximated to the midline to cover part of vocal folds, but did not contact each other. It did not produce mucosal wave. The vocal folds produced mucosal wave to form a single vibration sound source. For overtone höömii, the anterior two-thirds of ventricular folds touched each other during pronunciation. The last one-third produced the mucosal wave. The vocal folds produced mucosal wave at the same time, which was a composite vibration (double oscillator) sound source mode. The Höömii form of singing, including mixed voices and multivoice, was related to the presence of dual vibration sound sources. Its high overtone form of singing (whistle höömii) was related to stenosis at the resonance chambers' initiation site (ventricular folds level). Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.

Top